Fix typo of VLLM_ASCEND_ENABLE_TOPK_OPTIMIZE (#1112)
### What this PR does / why we need it? Fix typo of VLLM_ASCEND_ENABLE_TOPK_OPTIMIZE ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? CI passed Signed-off-by: linfeng-yuan <1102311262@qq.com>
This commit is contained in:
@@ -61,7 +61,7 @@ def test_models_distributed_DeepSeek():
|
||||
vllm_model.generate_greedy(example_prompts, max_tokens)
|
||||
|
||||
|
||||
@patch.dict(os.environ, {"VLLM_ASCEND_ENABLE_TOPK_OPTIMZE": "1"})
|
||||
@patch.dict(os.environ, {"VLLM_ASCEND_ENABLE_TOPK_OPTIMIZE": "1"})
|
||||
def test_models_distributed_topk() -> None:
|
||||
example_prompts = [
|
||||
"vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs.",
|
||||
|
||||
@@ -83,7 +83,7 @@ def test_multimodal(model, prompt_template, vllm_runner):
|
||||
max_tokens=64)
|
||||
|
||||
|
||||
@patch.dict(os.environ, {"VLLM_ASCEND_ENABLE_TOPK_OPTIMZE": "1"})
|
||||
@patch.dict(os.environ, {"VLLM_ASCEND_ENABLE_TOPK_OPTIMIZE": "1"})
|
||||
def test_models_topk() -> None:
|
||||
example_prompts = [
|
||||
"Hello, my name is",
|
||||
|
||||
Reference in New Issue
Block a user