[CI] remove old quantization model (#1003)

remove old quantization model, and new models will be added to testcase
later.

Signed-off-by: 22dimensions <waitingwind@foxmail.com>
This commit is contained in:
22dimensions
2025-06-10 10:07:36 +08:00
committed by GitHub
parent 706de02317
commit 5cd5d64242
2 changed files with 28 additions and 1 deletions

View File

@@ -78,6 +78,7 @@ class VllmRunner:
enable_chunked_prefill: bool = False,
swap_space: int = 4,
enforce_eager: Optional[bool] = True,
quantization: Optional[str] = None,
**kwargs,
) -> None:
self.model = LLM(
@@ -94,6 +95,7 @@ class VllmRunner:
max_model_len=max_model_len,
block_size=block_size,
enable_chunked_prefill=enable_chunked_prefill,
quantization=quantization,
**kwargs,
)