From 9830f85c425c7ff23236272890cbe5aca9d3184e Mon Sep 17 00:00:00 2001 From: Li Wang Date: Tue, 21 Oct 2025 10:31:55 +0800 Subject: [PATCH] [CI] Fix test_mla_v1 (#3570) ### What this PR does / why we need it? Remove test cases containing CPU incompatible operators ### Does this PR introduce _any_ user-facing change? ### How was this patch tested? - vLLM version: v0.11.0rc3 - vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0 Signed-off-by: wangli --- tests/ut/attention/test_mla_v1.py | 15 --------------- 1 file changed, 15 deletions(-) diff --git a/tests/ut/attention/test_mla_v1.py b/tests/ut/attention/test_mla_v1.py index 5e492e9..ff814c0 100644 --- a/tests/ut/attention/test_mla_v1.py +++ b/tests/ut/attention/test_mla_v1.py @@ -343,21 +343,6 @@ class TestAscendMLAImpl(TestBase): self.assertEqual(self.impl.num_queries_per_kv, 32) self.assertEqual(self.impl.tp_size, 2) - def test_v_up_proj(self): - batch_size = 4 - x = torch.randn(batch_size, self.impl.num_heads, - self.impl.kv_lora_rank) - - if not hasattr(self.impl, 'W_UV') or self.impl.W_UV is None: - self.impl.W_UV = torch.randn(self.impl.num_heads, - self.impl.kv_lora_rank, - self.impl.v_head_dim) - result = self.impl._v_up_proj(x) - - self.assertEqual(result.shape[0], batch_size) - self.assertEqual(result.shape[1], - self.impl.num_heads * self.impl.v_head_dim) - def test_q_proj_and_k_up_proj(self): batch_size = 4 x = torch.randn(batch_size, self.impl.num_heads, self.impl.qk_head_dim)