diff --git a/.github/workflows/vllm_ascend_test.yaml b/.github/workflows/vllm_ascend_test.yaml index 60f04d3..4aceed5 100644 --- a/.github/workflows/vllm_ascend_test.yaml +++ b/.github/workflows/vllm_ascend_test.yaml @@ -121,7 +121,6 @@ jobs: export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/local/Ascend/ascend-toolkit/latest/x86_64-linux/devlib pytest -sv --cov --cov-report=xml:unittests-coverage.xml tests/ut \ --ignore=tests/ut/test_platform.py \ - --ignore=tests/ut/patch/worker/patch_common/test_patch_minicpm.py \ --ignore=tests/ut/core/test_scheduler.py \ --ignore=tests/ut/kv_connector/test_llmdatadist_connector.py \ --ignore=tests/ut/kv_connector/test_mooncake_connector.py \ diff --git a/vllm_ascend/patch/worker/patch_common/__init__.py b/vllm_ascend/patch/worker/patch_common/__init__.py index 2e215b8..bed7e92 100644 --- a/vllm_ascend/patch/worker/patch_common/__init__.py +++ b/vllm_ascend/patch/worker/patch_common/__init__.py @@ -26,6 +26,4 @@ import vllm_ascend.patch.worker.patch_common.patch_logits # noqa import vllm_ascend.patch.worker.patch_common.patch_roberta # noqa import vllm_ascend.patch.worker.patch_common.patch_weight_loader # noqa import vllm_ascend.patch.worker.patch_common.patch_multimodal_merge # noqa - -# TODO: revert me when triton import is fixed -# import vllm_ascend.patch.worker.patch_common.patch_minicpm # noqa +import vllm_ascend.patch.worker.patch_common.patch_minicpm # noqa diff --git a/vllm_ascend/worker/model_runner_v1.py b/vllm_ascend/worker/model_runner_v1.py index 946c8ee..06c52bf 100644 --- a/vllm_ascend/worker/model_runner_v1.py +++ b/vllm_ascend/worker/model_runner_v1.py @@ -1346,6 +1346,8 @@ class NPUModelRunner(LoRAModelRunnerMixin): positions_cpu = self.positions_cpu[:num_input_tokens] positions = self.positions[:num_input_tokens] seq_lens_cpu = self.seq_lens_cpu[:num_reqs] + attn_state = self._build_attn_state(num_reqs, num_scheduled_tokens, + num_valid_tokens) self.attn_mask = self._make_attention_mask(seq_lens=seq_lens_cpu, position=positions_cpu, attn_state=attn_state)