diff --git a/vllm_ascend/attention/mla_v1.py b/vllm_ascend/attention/mla_v1.py index cb15bd1..73cbae6 100644 --- a/vllm_ascend/attention/mla_v1.py +++ b/vllm_ascend/attention/mla_v1.py @@ -202,6 +202,8 @@ class AscendMLAMetadataBuilder: npu_fused_infer_attention_score TND layout's limit of 16, \ got {self.decode_threshold}" + self.reorder_batch_threshold = self.decode_threshold + if self.chunked_prefill_enabled: self.chunked_prefill_workspace_size = min( # Max sure there is enough for 8 full length request or at least diff --git a/vllm_ascend/spec_decode/mtp_proposer.py b/vllm_ascend/spec_decode/mtp_proposer.py index ac0b3c5..ed4e887 100644 --- a/vllm_ascend/spec_decode/mtp_proposer.py +++ b/vllm_ascend/spec_decode/mtp_proposer.py @@ -555,7 +555,7 @@ class MtpProposer(Proposer): # copy inputs to buffer for cudagraph self.input_ids[:batch_size] = input_ids self.positions[:batch_size] = clamped_positions - self.hidden_states[:batch_size] = hidden_states + self.hidden_states[:hidden_states.shape[0]] = hidden_states attn_metadata_i.slot_mapping[:batch_size] = slot_mapping if attn_metadata_i.prefill is not None: