Fix potential flakiness in test_lora_qwen3 (#10250)

This commit is contained in:
Lifu Huang
2025-09-10 01:04:39 -07:00
committed by GitHub
parent 27760fc1b6
commit e903f695c8
3 changed files with 21 additions and 14 deletions

View File

@@ -24,6 +24,7 @@ from utils import (
CI_MULTI_LORA_MODELS,
TORCH_DTYPES,
LoRAModelCase,
ensure_reproducibility,
)
from sglang.test.runners import HFRunner, SRTRunner
@@ -76,13 +77,6 @@ class TestLoRA(CustomTestCase):
return batches
def ensure_reproducibility(self):
seed = 42
random.seed(seed)
torch.manual_seed(seed)
torch.cuda.manual_seed_all(seed)
torch.use_deterministic_algorithms(True)
def _run_lora_multiple_batch_on_model_cases(self, model_cases: List[LoRAModelCase]):
for model_case in model_cases:
for torch_dtype in TORCH_DTYPES:
@@ -121,14 +115,14 @@ class TestLoRA(CustomTestCase):
f"\n--- Running Batch {i} --- prompts: {prompts}, lora_paths: {lora_paths}"
)
self.ensure_reproducibility()
ensure_reproducibility()
srt_outputs = srt_runner.batch_forward(
prompts,
max_new_tokens=max_new_tokens,
lora_paths=lora_paths,
)
self.ensure_reproducibility()
ensure_reproducibility()
hf_outputs = hf_runner.forward(
prompts,
max_new_tokens=max_new_tokens,