model: support dots.vlm1 model (#8778)

Co-authored-by: weishi <bushou@xiaohongshu.com>
Co-authored-by: Ezra-Yu <1105212286@qq.com>
Co-authored-by: Jianfei Wang <905787410@qq.com>
Co-authored-by: qianwu <wangjianfei@xiaohongshu.com>
This commit is contained in:
chenge@xiaohongshu.com
2025-09-12 17:38:38 +08:00
committed by GitHub
parent 6d40308905
commit 1b1701f1f7
11 changed files with 806 additions and 11 deletions

View File

@@ -124,7 +124,9 @@ async def eval_mmmu(args) -> None:
answer_dict = {}
out_samples = {}
client = openai.AsyncOpenAI(
api_key="sk", base_url=f"http://127.0.0.1:{args.port}/v1"
api_key="sk",
base_url=f"http://127.0.0.1:{args.port}/v1",
timeout=20 * 60 * 60,
)
start = time.perf_counter()
base_url = f"http://127.0.0.1:{args.port}"
@@ -146,13 +148,14 @@ async def eval_mmmu(args) -> None:
_, response = await process_sample(
client, sample, sampling_params, lora_path
)
sample["original_response"] = response
answer = (
re.search(args.response_answer_regex, response)
if response is not None
else None
)
process_result(
answer.group(1) if answer else response,
answer.group(1).strip() if answer else response,
sample,
answer_dict,
out_samples,
@@ -168,13 +171,14 @@ async def eval_mmmu(args) -> None:
for coro in tqdm(asyncio.as_completed(tasks), total=len(tasks)):
sample, response = await coro
sample["original_response"] = response
answer = (
re.search(args.response_answer_regex, response)
if response is not None
else None
)
process_result(
answer.group(1) if answer else response,
answer.group(1).strip() if answer else response,
sample,
answer_dict,
out_samples,