[CI] Migrate e2e test runner to hk (#5344)

### What this PR does / why we need it?
This patch add new runner labels for the HK region, and e2e single-card
testing has been migrated to this runner.

- vLLM version: release/v0.13.0
- vLLM main:
bc0a5a0c08

---------

Signed-off-by: wangli <wangli858794774@gmail.com>
This commit is contained in:
Li Wang
2026-01-26 09:00:51 +08:00
committed by GitHub
parent 99bdd7363c
commit ca297eb57f
8 changed files with 40 additions and 23 deletions

View File

@@ -3,8 +3,8 @@
import vllm
import vllm.config
from modelscope import snapshot_download # type: ignore
from vllm.lora.request import LoRARequest
from unittest.mock import patch
from tests.e2e.conftest import VllmRunner
from vllm_ascend.utils import enable_custom_op
@@ -29,8 +29,8 @@ EXPECTED_LORA_OUTPUT = [
"SELECT poll_source FROM candidate GROUP BY poll_source ORDER BY count(*) DESC LIMIT 1", # noqa: E501
"SELECT poll_source FROM candidate GROUP BY poll_source ORDER BY count(*) DESC LIMIT 1", # noqa: E501
]
MODEL_PATH = "vllm-ascend/Llama-3.2-3B-Instruct"
# For hk region, we need to use the model from hf to avoid the network issue
MODEL_PATH = "meta-llama/Llama-3.2-3B-Instruct"
def do_sample(
@@ -105,9 +105,10 @@ def generate_and_test(llm,
print("removing lora")
@patch.dict("os.environ", {"VLLM_USE_MODELSCOPE": "False"})
def test_llama_lora(llama32_lora_files):
vllm_model = VllmRunner(
snapshot_download(MODEL_PATH),
MODEL_PATH,
enable_lora=True,
# also test odd max_num_seqs
max_num_seqs=7,