|
|
|
|
|
import os |
|
|
import sys |
|
|
import json |
|
|
import time |
|
|
import asyncio |
|
|
import logging |
|
|
from pathlib import Path |
|
|
from tqdm.asyncio import tqdm |
|
|
from argparse import ArgumentParser |
|
|
from playwright.async_api import async_playwright |
|
|
|
|
|
cmd_args = True |
|
|
parser = ArgumentParser() |
|
|
parser.add_argument('--test_dir', default='./test_webpages', help='the directory of test webpages.') |
|
|
parser.add_argument('--inference_dir', default='./inference', help='the directory of model output webpages.') |
|
|
parser.add_argument('--save_dir', default='./save_results', help='the directory for saving result info jsonl file.') |
|
|
parser.add_argument('--model_name', default="Qwen2.5-VL-32B-Instruct", help='using the vlms for your inference') |
|
|
parser.add_argument('--num_workers', default=50, help='num workers for computing') |
|
|
parser.add_argument('--log_dir', default='./', help='num workers for computing') |
|
|
if not cmd_args: |
|
|
args = parser.parse_args([]) |
|
|
else: |
|
|
args = parser.parse_args() |
|
|
|
|
|
MODEL_NAME = os.path.basename(args.model_name) |
|
|
LOG_PATH = os.path.join(args.log_dir, f'get_evaluation_{MODEL_NAME}.log') |
|
|
INFERENCE_DIR = args.inference_dir |
|
|
ORI_DIR = args.test_dir |
|
|
SAVE_PATH = os.path.join(args.save_dir, {MODEL_NAME}.jsonl) |
|
|
|
|
|
|
|
|
with open("./scripts/js/vue.global.js", "r", encoding="utf-8") as f: |
|
|
vue_code = f.read() |
|
|
with open("./scripts/js/one-color-all.js", "r", encoding="utf-8") as f: |
|
|
one_color_code = f.read() |
|
|
with open("./scripts/js/codeSim.js", "r", encoding="utf-8") as f: |
|
|
codesim_code = f.read() |
|
|
|
|
|
|
|
|
NUM_WORKERS = args.num_workers |
|
|
|
|
|
|
|
|
def setup_logger(is_console_handler=True): |
|
|
logger = logging.getLogger('web_scraping') |
|
|
logger.setLevel(logging.INFO) |
|
|
|
|
|
file_handler = logging.FileHandler(LOG_PATH) |
|
|
file_handler.setLevel(logging.INFO) |
|
|
|
|
|
if is_console_handler: |
|
|
console_handler = logging.StreamHandler(sys.stdout) |
|
|
console_handler.setLevel(logging.INFO) |
|
|
|
|
|
formatter = logging.Formatter('%(asctime)s - %(levelname)s - %(message)s', |
|
|
datefmt='%Y-%m-%d %H:%M:%S') |
|
|
file_handler.setFormatter(formatter) |
|
|
if is_console_handler: |
|
|
console_handler.setFormatter(formatter) |
|
|
|
|
|
logger.addHandler(file_handler) |
|
|
if is_console_handler: |
|
|
logger.addHandler(console_handler) |
|
|
|
|
|
return logger |
|
|
|
|
|
|
|
|
cache = [] |
|
|
file_lock = asyncio.Lock() |
|
|
|
|
|
|
|
|
async def save_result(name, layoutSim, styleSim, force=False): |
|
|
if force: |
|
|
async with file_lock: |
|
|
with open(SAVE_PATH, 'a') as f: |
|
|
for c in cache: |
|
|
f.write(json.dumps(c) + '\n') |
|
|
cache.clear() |
|
|
return 0 |
|
|
save_item = { |
|
|
'name': name, |
|
|
'groupLayoutScore': layoutSim['groupLayoutScore'], |
|
|
'overallScore': layoutSim['overallScore'], |
|
|
'relativeLayoutScore': layoutSim['relativeLayoutScore'], |
|
|
'relativeStyleScore': styleSim['relativeStyleScore'] |
|
|
} |
|
|
cache.append(save_item) |
|
|
if len(cache) >= NUM_WORKERS: |
|
|
async with file_lock: |
|
|
with open(SAVE_PATH, 'a') as f: |
|
|
for c in cache: |
|
|
f.write(json.dumps(c) + '\n') |
|
|
cache.clear() |
|
|
|
|
|
|
|
|
async def worker(name, queue, browser, logger, pbar): |
|
|
while not queue.empty(): |
|
|
html_item = await queue.get() |
|
|
name = os.path.splitext(html_item)[0] |
|
|
source_html_path = os.path.join(INFERENCE_DIR, f"{name}.html") |
|
|
target_html_path = os.path.join(ORI_DIR, name, "index.html") |
|
|
if not os.path.exists(source_html_path): |
|
|
logger.error(f"❌ File not Exsits: {source_html_path}") |
|
|
queue.task_done() |
|
|
pbar.update(1) |
|
|
continue |
|
|
if not os.path.exists(target_html_path): |
|
|
logger.error(f"❌ File not Exsits: {target_html_path}") |
|
|
queue.task_done() |
|
|
pbar.update(1) |
|
|
continue |
|
|
logger.info( |
|
|
f"⭐ push {source_html_path}") |
|
|
file_url = Path(source_html_path).as_uri() |
|
|
|
|
|
start_time = time.perf_counter() |
|
|
page = await browser.new_page(viewport={'width': 1920, 'height': 1080}) |
|
|
try: |
|
|
try: |
|
|
await page.goto( |
|
|
file_url, |
|
|
timeout=10000, |
|
|
wait_until='domcontentloaded' |
|
|
) |
|
|
except Exception as e: |
|
|
logger.error(f"⚠ Loading: {file_url}, Error: {e}") |
|
|
await page.add_script_tag(content=vue_code) |
|
|
await page.add_script_tag(content=one_color_code) |
|
|
await page.add_script_tag(content=codesim_code) |
|
|
sources = await page.evaluate("() => getElements()") |
|
|
elapsed = time.perf_counter() - start_time |
|
|
logger.info( |
|
|
f"✅ Source computed complete {source_html_path}, element count: {len(sources)} time: {elapsed:.2f}") |
|
|
except Exception as e: |
|
|
logger.error(f"❌ Source computed failed: {source_html_path}, Error: {e}") |
|
|
finally: |
|
|
await page.close() |
|
|
|
|
|
file_url = Path(target_html_path).as_uri() |
|
|
sec_start_time = time.perf_counter() |
|
|
page = await browser.new_page(viewport={'width': 1920, 'height': 1080}) |
|
|
|
|
|
|
|
|
try: |
|
|
try: |
|
|
await page.goto( |
|
|
file_url, |
|
|
timeout=3000, |
|
|
wait_until='domcontentloaded' |
|
|
) |
|
|
except Exception as e: |
|
|
logger.error(f"⚠ Loading: {file_url}, Error: {e}") |
|
|
await page.add_script_tag(content=vue_code) |
|
|
await page.add_script_tag(content=one_color_code) |
|
|
await page.add_script_tag(content=codesim_code) |
|
|
await page.evaluate("() => targetEls.value = getElements(false)") |
|
|
layoutSim = await page.evaluate("(sources) => getLayoutSim(sources)", sources) |
|
|
styleSim = await page.evaluate("(sources) => getStyleSim(sources)", sources) |
|
|
await save_result(name, layoutSim, styleSim) |
|
|
elapsed_total = time.perf_counter() - start_time |
|
|
elapsed = time.perf_counter() - sec_start_time |
|
|
logger.info( |
|
|
f"✅ Target computed complete {target_html_path}, layoutSim: {layoutSim['overallScore']}, styleSim: {styleSim['relativeStyleScore']}, element_count: {len(sources)} time: {elapsed:.2f} total_time: {elapsed_total:.2f}") |
|
|
except Exception as e: |
|
|
logger.error(f"❌ Target computed failed: {target_html_path}, Error: {e}") |
|
|
finally: |
|
|
await page.close() |
|
|
queue.task_done() |
|
|
pbar.update(1) |
|
|
|
|
|
|
|
|
async def main(): |
|
|
|
|
|
logger = setup_logger(False) |
|
|
source_list = os.listdir(INFERENCE_DIR) |
|
|
|
|
|
len_total = len(source_list) |
|
|
print('Total: ', len_total) |
|
|
exists_meta = [] |
|
|
if os.path.exists(SAVE_PATH): |
|
|
with open(SAVE_PATH) as f: |
|
|
exists_data = f.readlines() |
|
|
exists_data = [json.loads(item) for item in exists_data] |
|
|
for meta_info in exists_data: |
|
|
name = meta_info.get("name", 'none') |
|
|
exists_meta.append(name + '.html') |
|
|
|
|
|
exists_meta = set(exists_meta) |
|
|
source_list = [h for h in source_list if h not in exists_meta] |
|
|
print( |
|
|
f"Found: {len(exists_meta)} Matched: {len_total - len(source_list)} Remain: {len(source_list)}") |
|
|
|
|
|
|
|
|
queue = asyncio.Queue() |
|
|
for html_item in source_list: |
|
|
await queue.put(html_item) |
|
|
|
|
|
async with async_playwright() as p: |
|
|
browser_list = [] |
|
|
for i in range(NUM_WORKERS): |
|
|
browser = await p.chromium.launch(headless=True, args=['--no-sandbox', '--disable-setuid-sandbox']) |
|
|
browser_list.append(browser) |
|
|
print('Browser Started') |
|
|
with tqdm(total=len(source_list), desc="Progress") as pbar: |
|
|
tasks = [] |
|
|
for i in range(NUM_WORKERS): |
|
|
browser = browser_list[i % len(browser_list)] |
|
|
tasks.append(asyncio.create_task( |
|
|
worker(f"worker-{i}", queue, browser, logger, pbar))) |
|
|
await queue.join() |
|
|
for t in tasks: |
|
|
t.cancel() |
|
|
for browser in browser_list: |
|
|
await browser.close() |
|
|
save_result(None, None, None, True) |
|
|
|
|
|
if __name__ == "__main__": |
|
|
asyncio.run(main()) |
|
|
|
|
|
|
|
|
|