[CI] fix port conflicts (#5789)
This commit is contained in:
@@ -14,7 +14,7 @@ class TestFile:
|
||||
suites = {
|
||||
"per-commit": [
|
||||
TestFile("models/lora/test_lora.py", 76),
|
||||
TestFile("models/lora/test_lora_backend.py", 420),
|
||||
TestFile("models/lora/test_lora_backend.py", 99),
|
||||
TestFile("models/lora/test_multi_lora_backend.py", 60),
|
||||
TestFile("models/test_embedding_models.py", 35),
|
||||
TestFile("models/test_generation_models.py", 103),
|
||||
@@ -23,30 +23,30 @@ suites = {
|
||||
TestFile("models/test_compressed_tensors_models.py", 100),
|
||||
TestFile("models/test_reward_models.py", 83),
|
||||
TestFile("models/test_gme_qwen_models.py", 45),
|
||||
TestFile("models/test_clip_models.py", 100),
|
||||
TestFile("models/test_vlm_models.py", 100),
|
||||
TestFile("models/test_clip_models.py", 52),
|
||||
TestFile("models/test_vlm_models.py", 581),
|
||||
TestFile("test_abort.py", 51),
|
||||
TestFile("test_block_int8.py", 22),
|
||||
TestFile("test_chunked_prefill.py", 336),
|
||||
TestFile("test_eagle_infer.py", 500),
|
||||
TestFile("test_chunked_prefill.py", 285),
|
||||
TestFile("test_eagle_infer.py", 584),
|
||||
TestFile("test_ebnf_constrained.py"),
|
||||
TestFile("test_fa3.py", 400),
|
||||
TestFile("test_fa3.py", 376),
|
||||
TestFile("test_fp8_kernel.py", 8),
|
||||
TestFile("test_embedding_openai_server.py", 36),
|
||||
TestFile("test_embedding_openai_server.py", 141),
|
||||
TestFile("test_hidden_states.py", 55),
|
||||
TestFile("test_int8_kernel.py", 8),
|
||||
TestFile("test_input_embeddings.py", 38),
|
||||
TestFile("test_json_constrained.py", 98),
|
||||
TestFile("test_large_max_new_tokens.py", 41),
|
||||
TestFile("test_metrics.py", 32),
|
||||
TestFile("test_mla.py", 162),
|
||||
TestFile("test_mla.py", 242),
|
||||
TestFile("test_mla_deepseek_v3.py", 221),
|
||||
TestFile("test_mla_int8_deepseek_v3.py", 522),
|
||||
TestFile("test_mla_int8_deepseek_v3.py", 674),
|
||||
TestFile("test_mla_flashinfer.py", 395),
|
||||
TestFile("test_mla_fp8.py", 93),
|
||||
TestFile("test_mla_fp8.py", 153),
|
||||
TestFile("test_no_chunked_prefill.py", 126),
|
||||
TestFile("test_no_overlap_scheduler.py", 262),
|
||||
TestFile("test_openai_server.py", 186),
|
||||
TestFile("test_openai_server.py", 149),
|
||||
TestFile("test_penalty.py", 41),
|
||||
TestFile("test_page_size.py", 60),
|
||||
TestFile("test_pytorch_sampling_backend.py", 66),
|
||||
@@ -57,11 +57,11 @@ suites = {
|
||||
TestFile("test_request_length_validation.py", 31),
|
||||
TestFile("test_retract_decode.py", 54),
|
||||
TestFile("test_server_args.py", 1),
|
||||
TestFile("test_skip_tokenizer_init.py", 72),
|
||||
TestFile("test_skip_tokenizer_init.py", 117),
|
||||
TestFile("test_srt_engine.py", 237),
|
||||
TestFile("test_srt_endpoint.py", 94),
|
||||
TestFile("test_torch_compile.py", 76),
|
||||
TestFile("test_torch_compile_moe.py", 85),
|
||||
TestFile("test_torch_compile_moe.py", 235),
|
||||
TestFile("test_torch_native_attention_backend.py", 123),
|
||||
TestFile("test_torchao.py", 70),
|
||||
TestFile("test_triton_attention_kernels.py", 4),
|
||||
@@ -69,27 +69,27 @@ suites = {
|
||||
TestFile("test_update_weights_from_disk.py", 114),
|
||||
TestFile("test_update_weights_from_tensor.py", 48),
|
||||
TestFile("test_vertex_endpoint.py", 31),
|
||||
TestFile("test_vision_chunked_prefill.py", 99),
|
||||
TestFile("test_vision_chunked_prefill.py", 119),
|
||||
TestFile("test_vlm_accuracy.py", 60),
|
||||
TestFile("test_vision_openai_server.py", 537),
|
||||
TestFile("test_vision_openai_server.py", 637),
|
||||
TestFile("test_fim_completion.py", 40),
|
||||
TestFile("test_w8a8_quantization.py", 46),
|
||||
TestFile("test_eval_fp8_accuracy.py", 303),
|
||||
TestFile("test_create_kvindices.py", 2),
|
||||
TestFile("test_hicache.py", 60),
|
||||
TestFile("test_hicache_mla.py", 90),
|
||||
TestFile("test_hicache.py", 116),
|
||||
TestFile("test_hicache_mla.py", 254),
|
||||
TestFile("test_fused_moe.py", 30),
|
||||
TestFile("test_triton_moe_channel_fp8_kernel.py", 25),
|
||||
],
|
||||
"per-commit-2-gpu": [
|
||||
TestFile("models/lora/test_lora_tp.py", 150),
|
||||
TestFile("test_data_parallelism.py", 90),
|
||||
TestFile("test_dp_attention.py", 150),
|
||||
TestFile("test_mla_tp.py", 174),
|
||||
TestFile("test_moe_ep.py", 220),
|
||||
TestFile("test_patch_torch.py", 30),
|
||||
TestFile("test_update_weights_from_distributed.py", 100),
|
||||
TestFile("test_verl_engine.py", 100),
|
||||
TestFile("models/lora/test_lora_tp.py", 116),
|
||||
TestFile("test_data_parallelism.py", 73),
|
||||
TestFile("test_dp_attention.py", 137),
|
||||
TestFile("test_mla_tp.py", 170),
|
||||
TestFile("test_moe_ep.py", 181),
|
||||
TestFile("test_patch_torch.py", 19),
|
||||
TestFile("test_update_weights_from_distributed.py", 103),
|
||||
TestFile("test_verl_engine.py", 64),
|
||||
],
|
||||
"per-commit-8-gpu": [
|
||||
TestFile("test_local_attn.py", 250),
|
||||
|
||||
Reference in New Issue
Block a user