From 66db0704237c7360da4f5c9271de6c2d0b60e1d3 Mon Sep 17 00:00:00 2001 From: ZT-AIA <63220130+ZT-AIA@users.noreply.github.com> Date: Tue, 31 Mar 2026 16:52:58 +0800 Subject: [PATCH] [cherry-pick][Test]repair for test_compute_slot_mapping (#7836) ### What this PR does / why we need it? repair for test_compute_slot_mapping Signed-off-by: ZT-AIA <1028681969@qq.com> --- .../ops/singlecard_ops/triton/test_compute_slot_mapping.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/tests/e2e/nightly/single_node/ops/singlecard_ops/triton/test_compute_slot_mapping.py b/tests/e2e/nightly/single_node/ops/singlecard_ops/triton/test_compute_slot_mapping.py index 6e23b982..a041a85a 100644 --- a/tests/e2e/nightly/single_node/ops/singlecard_ops/triton/test_compute_slot_mapping.py +++ b/tests/e2e/nightly/single_node/ops/singlecard_ops/triton/test_compute_slot_mapping.py @@ -5,7 +5,6 @@ from vllm.v1.worker.gpu.block_table import _compute_slot_mappings_kernel as \ ref_compute_slot_mappings_kernel from vllm_ascend.worker.v2.block_table import _compute_slot_mappings_kernel as \ ascend_compute_slot_mappings_kernel -from vllm.v1.worker.gpu.block_table import _load_ptr, _make_ptr_tensor def test_compute_slot_mapping_npu_kernel(): @@ -52,7 +51,7 @@ def test_compute_slot_mapping_npu_kernel(): for i in range(num_kv_cache_groups): block_table = torch.randint(0, 320, (max_num_reqs, max_num_blocks), dtype=torch.int32, device=device) block_tables.append(block_table) - block_table_ptrs = _make_ptr_tensor(block_tables) + block_table_ptrs = torch.tensor([t.data_ptr() for t in block_table], dtype=torch.uint64, device=device) block_table_strides = torch.tensor([320], dtype=torch.int32, device=device) block_sizes_tensor = torch.tensor([128], dtype=torch.int32, device=device)