Fix flaky CI: test_vlm_models (#8006)

This commit is contained in:
Lifu Huang
2025-07-14 14:56:41 -07:00
committed by GitHub
parent 1ebec1a8b0
commit d969504d9a
2 changed files with 0 additions and 4 deletions

View File

@@ -76,7 +76,6 @@ class MultiModalityDataPaddingPatternTokenPairs(MultiModalityDataPaddingPattern)
This function will replace the data-tokens in between with pad_values accordingly
"""
pad_values = [item.pad_value for item in mm_inputs.mm_items]
print(f"{mm_inputs.mm_items=}")
data_token_pairs = self.data_token_id_pairs
mm_inputs.data_offsets = []
if data_token_pairs is None:

View File

@@ -36,7 +36,6 @@ class Gemma3SGLangImageProcessor(SGLangBaseProcessor):
*args,
**kwargs,
):
print(f"{image_data=}")
base_output = self.load_mm_data(
prompt=input_text,
image_data=image_data,
@@ -48,8 +47,6 @@ class Gemma3SGLangImageProcessor(SGLangBaseProcessor):
)
mm_items, input_ids, _ = self.process_and_combine_mm_data(base_output)
print(f"{base_output=}")
print(f"{mm_items=}")
return {
"input_ids": input_ids.tolist(),
"mm_items": mm_items,