Revert "Set csgmv as default lora backend. (#11488)" (#11735)

This commit is contained in:
Yineng Zhang
2025-10-17 10:01:36 -07:00
committed by GitHub
parent 9b0f725b1d
commit da681f35d3
11 changed files with 23 additions and 11 deletions

View File

@@ -71,6 +71,7 @@ class TestLoRAQwen3(CustomTestCase):
for model_case in model_cases:
for torch_dtype in TORCH_DTYPES:
max_new_tokens = 32
backend = "triton"
base_path = model_case.base
lora_adapter_paths = [a.name for a in model_case.adaptors]
assert len(lora_adapter_paths) >= 2
@@ -127,7 +128,7 @@ class TestLoRAQwen3(CustomTestCase):
]
print(
f"\n========== Testing multiple batches on base '{base_path}', dtype={torch_dtype} ---"
f"\n========== Testing multiple batches on base '{base_path}' with backend={backend}, dtype={torch_dtype} ---"
)
# Initialize runners
@@ -138,6 +139,7 @@ class TestLoRAQwen3(CustomTestCase):
model_type="generation",
lora_paths=[lora_adapter_paths[0], lora_adapter_paths[1]],
max_loras_per_batch=len(lora_adapter_paths) + 1,
lora_backend=backend,
sleep_on_idle=True, # Eliminate non-determinism by forcing all requests to be processed in one batch.
attention_backend="torch_native",
)
@@ -181,7 +183,7 @@ class TestLoRAQwen3(CustomTestCase):
if rouge_score < rouge_tol:
raise AssertionError(
f"ROUGE-L score {rouge_score} below tolerance {rouge_tol} "
f"for base '{base_path}', adaptor '{lora_paths}', prompt: '{prompts}...'"
f"for base '{base_path}', adaptor '{lora_paths}', backend '{backend}', prompt: '{prompts}...'"
)
print(f"--- Batch {i+1} Comparison Passed --- ")