[2/N][refactor] torchair deepseek mla backend refactor (#2459)
### What this PR does / why we need it?
This PR move current unified mla backend to torchair folder and remove
torchair-related code in attention/mla_v1.py (1.3k -> 0.9k).
### Does this PR introduce _any_ user-facing change?
No.
### How was this patch tested?
Running eager mode with mla backend, and torchair mode with code before
[2445](https://github.com/vllm-project/vllm-ascend/pull/2445)
- vLLM version: v0.10.0
- vLLM main:
f571ff8eb6
Signed-off-by: linfeng-yuan <1102311262@qq.com>
This commit is contained in:
@@ -1,14 +1,12 @@
|
||||
from dataclasses import dataclass
|
||||
from typing import TYPE_CHECKING, Optional, Tuple, Type, TypeVar
|
||||
|
||||
import numpy as np
|
||||
import torch
|
||||
import torch.nn as nn
|
||||
import torch_npu
|
||||
from vllm.attention.backends.abstract import (AttentionBackend, AttentionLayer,
|
||||
AttentionMetadata,
|
||||
MLAAttentionImpl)
|
||||
from vllm.attention.backends.utils import PAD_SLOT_ID
|
||||
from vllm.config import VllmConfig, get_current_vllm_config
|
||||
from vllm.distributed import get_tensor_model_parallel_world_size
|
||||
from vllm.model_executor.layers.linear import (LinearBase,
|
||||
@@ -24,9 +22,6 @@ from vllm_ascend.multistream.base import MSAttentionMetadataSplitConfig
|
||||
from vllm_ascend.multistream.context import get_multistream_comm_context
|
||||
from vllm_ascend.multistream.ms_split import model_input_split_v1_mla_attn
|
||||
from vllm_ascend.ops.attention import vanilla_chunked_prefill_mla
|
||||
from vllm_ascend.torchair.utils import (TorchairCommonAttentionMetadata,
|
||||
npu_stream_switch, npu_wait_tensor)
|
||||
from vllm_ascend.utils import npu_prefetch
|
||||
from vllm_ascend.worker.npu_input_batch import InputBatch
|
||||
|
||||
if TYPE_CHECKING:
|
||||
@@ -212,8 +207,6 @@ class AscendMLAMetadataBuilder:
|
||||
dtype=self.model_config.dtype,
|
||||
device=device,
|
||||
)
|
||||
ascend_config = get_ascend_config()
|
||||
self.torchair_graph_enabled = ascend_config.torchair_graph_config.enabled
|
||||
self.rope_dim = self.model_config.hf_text_config.qk_rope_head_dim
|
||||
self.cos_cache = None
|
||||
self.sin_cache = None
|
||||
@@ -231,20 +224,10 @@ class AscendMLAMetadataBuilder:
|
||||
|
||||
for i, req_id in enumerate(input_batch.req_ids):
|
||||
num_tokens = scheduler_output.num_scheduled_tokens[req_id]
|
||||
num_spec_tokens = len(
|
||||
scheduler_output.scheduled_spec_decode_tokens.get(req_id, []))
|
||||
# For torch air graph mode we treat spec decoding as decode.
|
||||
if self.torchair_graph_enabled:
|
||||
if num_tokens - num_spec_tokens == 1:
|
||||
decodes.append(i)
|
||||
else:
|
||||
prefills.append(i)
|
||||
# For eager mode we treat spec decoding as chunked prefill.
|
||||
if num_tokens == 1:
|
||||
decodes.append(i)
|
||||
else:
|
||||
if num_tokens == 1:
|
||||
decodes.append(i)
|
||||
else:
|
||||
prefills.append(i)
|
||||
prefills.append(i)
|
||||
|
||||
# We hope that this is fairly minimal since decodes
|
||||
# should be around for a number of iterations so hopefully they are
|
||||
@@ -277,99 +260,6 @@ class AscendMLAMetadataBuilder:
|
||||
# better way of doing this
|
||||
return modified_batch
|
||||
|
||||
def _get_graph_runner_block_tables(
|
||||
self, num_seqs: int, block_tables: torch.Tensor) -> torch.Tensor:
|
||||
max_blocks = self.max_blocks
|
||||
|
||||
graph_block_tables = torch.zeros((num_seqs, max_blocks),
|
||||
dtype=block_tables.dtype,
|
||||
device=block_tables.device)
|
||||
|
||||
num_blocks = block_tables.size(1)
|
||||
if num_blocks <= max_blocks:
|
||||
graph_block_tables[:num_seqs, :
|
||||
num_blocks] = block_tables[:num_seqs, :
|
||||
num_blocks]
|
||||
else:
|
||||
graph_block_tables[:num_seqs, :
|
||||
max_blocks] = block_tables[:num_seqs, :
|
||||
max_blocks]
|
||||
|
||||
return graph_block_tables[:, :max_blocks]
|
||||
|
||||
def build_torchair_graph_dummy(
|
||||
self,
|
||||
common_attn_metadata: TorchairCommonAttentionMetadata,
|
||||
) -> AscendMLAMetadata:
|
||||
device = self.device
|
||||
num_reqs = common_attn_metadata.num_reqs
|
||||
block_table = torch.zeros((num_reqs, self.max_blocks),
|
||||
dtype=torch.int32,
|
||||
device=device)
|
||||
block_table = self._get_graph_runner_block_tables(
|
||||
num_reqs, block_table)
|
||||
num_tokens = num_reqs * common_attn_metadata.decode_token_per_req
|
||||
seq_lens = torch.zeros(num_reqs, dtype=torch.int32, device=device)
|
||||
seq_lens_list = [0] * num_reqs
|
||||
input_positions = torch.zeros(num_tokens,
|
||||
dtype=torch.int32,
|
||||
device=device).long()
|
||||
slot_mapping = torch.full((num_tokens, ),
|
||||
PAD_SLOT_ID,
|
||||
dtype=torch.int32,
|
||||
device=device)
|
||||
query_start_loc = torch.full((num_reqs, ),
|
||||
-1,
|
||||
dtype=torch.int32,
|
||||
device=device)
|
||||
sin = torch.ones(num_tokens,
|
||||
1,
|
||||
1,
|
||||
self.rope_dim,
|
||||
dtype=self.model_config.dtype,
|
||||
device=device)
|
||||
cos = torch.ones(num_tokens,
|
||||
1,
|
||||
1,
|
||||
self.rope_dim,
|
||||
dtype=self.model_config.dtype,
|
||||
device=device)
|
||||
if self.vllm_config.speculative_config is not None and\
|
||||
self.vllm_config.speculative_config.method == 'deepseek_mtp':
|
||||
attn_state = AscendAttentionState.SpecDecoding
|
||||
num_decode_tokens = 2
|
||||
else:
|
||||
attn_state = AscendAttentionState.DecodeOnly
|
||||
num_decode_tokens = 1
|
||||
decode_metadata = AscendMLADecodeMetadata(
|
||||
input_positions=input_positions,
|
||||
block_table=block_table,
|
||||
seq_lens=seq_lens,
|
||||
seq_lens_list=seq_lens_list,
|
||||
max_seq_lens=1,
|
||||
attn_mask=common_attn_metadata.spec_attn_mask,
|
||||
actual_seq_lengths_q=common_attn_metadata.
|
||||
actual_seq_lengths_q[:num_reqs],
|
||||
sin=sin,
|
||||
cos=cos,
|
||||
)
|
||||
return self.metadata_cls( # type: ignore
|
||||
num_input_tokens=common_attn_metadata.num_actual_tokens,
|
||||
num_actual_tokens=common_attn_metadata.num_actual_tokens,
|
||||
slot_mapping=slot_mapping,
|
||||
head_dim=self.model_config.get_head_size(),
|
||||
num_decodes=1,
|
||||
num_decode_tokens=num_decode_tokens,
|
||||
num_prefills=0,
|
||||
attn_mask=common_attn_metadata.attn_mask,
|
||||
attn_state=attn_state,
|
||||
prefill=None,
|
||||
decode=decode_metadata,
|
||||
query_start_loc=query_start_loc,
|
||||
seq_lens=seq_lens,
|
||||
block_tables=block_table,
|
||||
)
|
||||
|
||||
def build(
|
||||
self,
|
||||
common_attn_metadata: AscendCommonAttentionMetadata,
|
||||
@@ -379,14 +269,8 @@ class AscendMLAMetadataBuilder:
|
||||
num_actual_tokens = common_attn_metadata.num_actual_tokens
|
||||
query_start_loc = common_attn_metadata.query_start_loc
|
||||
query_start_loc_cpu = common_attn_metadata.query_start_loc_cpu
|
||||
if self.torchair_graph_enabled and common_attn_metadata.attn_state in [
|
||||
AscendAttentionState.DecodeOnly,
|
||||
AscendAttentionState.SpecDecoding
|
||||
]:
|
||||
decode_threshold = common_attn_metadata.decode_token_per_req
|
||||
else:
|
||||
# TODO(xyx): remove the if condition after mla supports torch mode speculative decoding
|
||||
decode_threshold = 1
|
||||
# TODO(xyx): remove the if condition after mla supports torch mode speculative decoding
|
||||
decode_threshold = 1
|
||||
num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = \
|
||||
split_decodes_and_prefills(common_attn_metadata, decode_threshold=decode_threshold)
|
||||
assert num_decodes + num_prefills == num_reqs
|
||||
@@ -489,57 +373,14 @@ class AscendMLAMetadataBuilder:
|
||||
)
|
||||
|
||||
decode_metadata = None
|
||||
graph_pad_size = common_attn_metadata.graph_pad_size
|
||||
use_torchair_graph = graph_pad_size != -1
|
||||
if num_decodes > 0:
|
||||
actual_seq_lengths_q = query_start_loc[1:].tolist()
|
||||
max_seq_lens = seq_lens[:num_decodes].max().item()
|
||||
seq_lens = seq_lens[:num_decode_tokens]
|
||||
input_positions = input_positions[:num_decode_tokens]
|
||||
block_table = block_table[:num_decode_tokens, ...]
|
||||
if use_torchair_graph and common_attn_metadata.attn_state in [
|
||||
AscendAttentionState.DecodeOnly,
|
||||
AscendAttentionState.SpecDecoding
|
||||
]:
|
||||
num_reqs_pad_size = 0
|
||||
num_token_pad_size = 0
|
||||
if graph_pad_size != 0:
|
||||
pad_value = 0
|
||||
num_token_pad_size = graph_pad_size - num_decode_tokens
|
||||
num_reqs_pad_size = (
|
||||
graph_pad_size //
|
||||
common_attn_metadata.decode_token_per_req - num_reqs)
|
||||
padded_seq_lens = seq_lens.tolist(
|
||||
) + [pad_value] * num_reqs_pad_size
|
||||
else:
|
||||
padded_seq_lens = seq_lens.tolist()
|
||||
|
||||
seq_lens = torch.from_numpy(
|
||||
np.array(padded_seq_lens).astype(np.int32))
|
||||
seq_lens_list = padded_seq_lens
|
||||
slot_padding = torch.full((num_token_pad_size, ),
|
||||
PAD_SLOT_ID,
|
||||
dtype=slot_mapping.dtype,
|
||||
device=slot_mapping.device)
|
||||
slot_mapping = torch.cat([slot_mapping, slot_padding])
|
||||
block_table_padding = torch.zeros(
|
||||
(num_reqs_pad_size, ) + block_table.shape[1:],
|
||||
dtype=block_table.dtype,
|
||||
device=block_table.device)
|
||||
block_table = torch.cat([block_table, block_table_padding],
|
||||
dim=0)
|
||||
block_table = self._get_graph_runner_block_tables(
|
||||
num_reqs + num_reqs_pad_size, block_table)
|
||||
position_padding = torch.zeros(num_token_pad_size,
|
||||
dtype=input_positions.dtype,
|
||||
device=input_positions.device)
|
||||
input_positions = torch.cat(
|
||||
[input_positions, position_padding])
|
||||
actual_seq_lengths_q = query_start_loc[1:].tolist(
|
||||
) + common_attn_metadata.actual_seq_lengths_q[
|
||||
num_reqs:num_reqs + num_reqs_pad_size]
|
||||
else:
|
||||
seq_lens_list = seq_lens.tolist()
|
||||
seq_lens_list = seq_lens.tolist()
|
||||
# TODO(xyx): whether this block is necessary without torchair
|
||||
# mtp torchair + PD scenario, last element of actual_seq_lengths_q must equal to batch_size(num_tokens)
|
||||
batch_size = slot_mapping.size(0)
|
||||
if actual_seq_lengths_q[-1] != batch_size \
|
||||
@@ -624,8 +465,6 @@ class AscendMLAImpl(MLAAttentionImpl):
|
||||
self.tp_size = get_tensor_model_parallel_world_size()
|
||||
|
||||
ascend_config = get_ascend_config()
|
||||
self.torchair_graph_enabled = ascend_config.torchair_graph_config.enabled
|
||||
self.enable_kv_nz = ascend_config.torchair_graph_config.enable_kv_nz
|
||||
self.enable_shared_expert_dp = ascend_config.enable_shared_expert_dp
|
||||
|
||||
# Adapt torch air graph mode with spec decoding.
|
||||
@@ -634,21 +473,14 @@ class AscendMLAImpl(MLAAttentionImpl):
|
||||
self.spec_token_num = speculative_config.num_speculative_tokens
|
||||
assert self.spec_token_num > 0
|
||||
|
||||
def _v_up_proj_and_o_proj(self, x, enable_multistream_mla: bool = False):
|
||||
def _v_up_proj_and_o_proj(self, x):
|
||||
# Convert from (B, N, L) to (N, B, L)
|
||||
x = x.view(-1, self.num_heads, self.kv_lora_rank).transpose(0, 1)
|
||||
# Multiply (N, B, L) x (N, L, V) -> (N, B, V)
|
||||
x = torch.bmm(x, self.W_UV)
|
||||
# Convert from (N, B, V) to (B, N * V)
|
||||
x = x.transpose(0, 1).reshape(-1, self.num_heads * self.v_head_dim)
|
||||
if hasattr(self, "running_in_graph") and not self.running_in_graph:
|
||||
return x
|
||||
MAX_O_PROJ_PREFETCH_SIZE = 16 * 1024 * 1024 # 16MB
|
||||
npu_prefetch(self.o_proj.weight,
|
||||
x,
|
||||
max_size=MAX_O_PROJ_PREFETCH_SIZE,
|
||||
enabled=enable_multistream_mla)
|
||||
return self.o_proj(x, is_prefill=False)[0]
|
||||
return x
|
||||
|
||||
# Return `ql_nope`, `q_pe`
|
||||
def _q_proj_and_k_up_proj(self, x):
|
||||
@@ -915,77 +747,6 @@ class AscendMLAImpl(MLAAttentionImpl):
|
||||
|
||||
return attn_output
|
||||
|
||||
def exec_kv(
|
||||
self,
|
||||
hidden_states: torch.Tensor,
|
||||
cos: torch.Tensor,
|
||||
sin: torch.Tensor,
|
||||
kv_cache: Tuple,
|
||||
slots: torch.Tensor,
|
||||
):
|
||||
|
||||
B = hidden_states.shape[0]
|
||||
N = self.num_kv_heads
|
||||
S = 1
|
||||
kv = self.kv_a_proj_with_mqa(hidden_states)[0]
|
||||
# npu_kv_rmsnorm_rope_cache needs [B, N, S, D]
|
||||
kv = kv.view(B, N, S, self.kv_lora_rank + self.qk_rope_head_dim)
|
||||
cache_mode = "PA_NZ" if self.enable_kv_nz else "PA"
|
||||
k_pe, k_nope, _, _ = torch_npu.npu_kv_rmsnorm_rope_cache(
|
||||
kv,
|
||||
self.kv_a_layernorm.weight,
|
||||
cos,
|
||||
sin,
|
||||
slots.to(torch.int64),
|
||||
kv_cache[1],
|
||||
kv_cache[0],
|
||||
epsilon=self.kv_a_layernorm.variance_epsilon,
|
||||
cache_mode=cache_mode,
|
||||
)
|
||||
return k_pe, k_nope, kv
|
||||
|
||||
def exec_kv_prefill(
|
||||
self,
|
||||
hidden_states: torch.Tensor,
|
||||
cos: torch.Tensor,
|
||||
sin: torch.Tensor,
|
||||
kv_cache: Tuple,
|
||||
slots: torch.Tensor,
|
||||
):
|
||||
|
||||
B = hidden_states.shape[0]
|
||||
N = self.num_kv_heads
|
||||
S = 1
|
||||
kv = self.kv_a_proj_with_mqa(hidden_states)[0]
|
||||
# npu_kv_rmsnorm_rope_cache needs [B, N, S, D]
|
||||
kv = kv.view(B, N, S, self.kv_lora_rank + self.qk_rope_head_dim)
|
||||
cache_mode = "PA_BLK_NZ" if self.enable_kv_nz else "PA"
|
||||
_, _, k_pe, k_nope = torch_npu.npu_kv_rmsnorm_rope_cache(
|
||||
kv,
|
||||
self.kv_a_layernorm.weight,
|
||||
cos,
|
||||
sin,
|
||||
slots.to(torch.int64),
|
||||
kv_cache[1],
|
||||
kv_cache[0],
|
||||
epsilon=self.kv_a_layernorm.variance_epsilon,
|
||||
cache_mode=cache_mode,
|
||||
is_output_kv=True,
|
||||
)
|
||||
return k_pe, k_nope
|
||||
|
||||
def rope_single(
|
||||
self,
|
||||
x: torch.Tensor,
|
||||
cos: torch.Tensor,
|
||||
sin: torch.Tensor,
|
||||
) -> torch.Tensor:
|
||||
B, N, D = x.shape
|
||||
S = 1
|
||||
x = x.view(B, N, S, D)
|
||||
x = torch_npu.npu_interleave_rope(x, cos, sin)
|
||||
return x.view(B, N, D)
|
||||
|
||||
def _forward_decode(
|
||||
self,
|
||||
q_nope: torch.Tensor,
|
||||
@@ -994,100 +755,41 @@ class AscendMLAImpl(MLAAttentionImpl):
|
||||
k_pe: torch.Tensor,
|
||||
kv_c_and_k_pe_cache: Tuple[torch.Tensor],
|
||||
attn_metadata: AscendMLAMetadata,
|
||||
enable_multistream_mla: bool = False,
|
||||
) -> torch.Tensor:
|
||||
decode_meta = attn_metadata.decode
|
||||
assert decode_meta is not None
|
||||
num_tokens = q_nope.size(0)
|
||||
if self.running_in_graph or self.running_chunkprefilll_with_torchair:
|
||||
# shape of knope/k_pe for npu graph mode should be:
|
||||
# [num_blocks, num_kv_heads, block_size, self.kv_lora_rank/self.qk_rope_head_dim]
|
||||
block_size = kv_c_and_k_pe_cache[0].shape[1]
|
||||
actual_seq_lengths = None
|
||||
if self.enable_kv_nz:
|
||||
k_nope = k_nope.view(-1, self.num_kv_heads,
|
||||
self.kv_lora_rank // 16, block_size, 16)
|
||||
k_pe = k_pe.view(-1, self.num_kv_heads,
|
||||
self.qk_rope_head_dim // 16, block_size, 16)
|
||||
input_layout = "BSND"
|
||||
else:
|
||||
k_nope = k_nope.view(-1, self.num_kv_heads, block_size,
|
||||
self.kv_lora_rank)
|
||||
k_pe = k_pe.view(-1, self.num_kv_heads, block_size,
|
||||
self.qk_rope_head_dim)
|
||||
input_layout = "BNSD"
|
||||
|
||||
if attn_metadata.attn_state == AscendAttentionState.SpecDecoding:
|
||||
assert num_tokens % self.spec_token_num == 0
|
||||
input_layout = "TND"
|
||||
# [bs * q_seq_len, num_heads_per_rank, dim]
|
||||
q_nope = q_nope.view(num_tokens, self.num_heads, -1)
|
||||
q_pe = q_pe.view(num_tokens, self.num_heads, -1)
|
||||
sparse_mode = 3
|
||||
spec_attn_mask = attn_metadata.decode.attn_mask # type:ignore
|
||||
actual_seq_lengths = decode_meta.actual_seq_lengths_q
|
||||
else:
|
||||
if self.enable_kv_nz:
|
||||
q_nope = q_nope.view(num_tokens, 1, self.num_heads, -1)
|
||||
q_pe = q_pe.view(num_tokens, 1, self.num_heads, -1)
|
||||
else:
|
||||
q_nope = q_nope.view(num_tokens, self.num_heads, 1, -1)
|
||||
q_pe = q_pe.view(num_tokens, self.num_heads, 1, -1)
|
||||
sparse_mode = 0
|
||||
spec_attn_mask = None
|
||||
|
||||
attn_output, _ = torch_npu.npu_fused_infer_attention_score(
|
||||
q_nope,
|
||||
k_nope,
|
||||
k_nope,
|
||||
query_rope=q_pe,
|
||||
key_rope=k_pe,
|
||||
num_heads=self.num_heads,
|
||||
num_key_value_heads=self.num_kv_heads,
|
||||
input_layout=input_layout,
|
||||
atten_mask=spec_attn_mask,
|
||||
sparse_mode=sparse_mode,
|
||||
scale=self.scale,
|
||||
antiquant_mode=0,
|
||||
antiquant_scale=None,
|
||||
block_table=decode_meta.block_table,
|
||||
block_size=block_size,
|
||||
actual_seq_lengths_kv=decode_meta.seq_lens_list,
|
||||
actual_seq_lengths=actual_seq_lengths)
|
||||
# The MLA_PA path will be used as default path in the future, `_npu_paged_attention_mla` will
|
||||
# be removed after the torch_npu contains `torch_npu.atb.npu_multi_head_latent_attention` become
|
||||
# public available
|
||||
assert len(kv_c_and_k_pe_cache) > 1
|
||||
if envs_ascend.VLLM_ASCEND_MLA_PA:
|
||||
attn_output = torch_npu.atb.npu_multi_head_latent_attention(
|
||||
q_nope, q_pe, kv_c_and_k_pe_cache[0], kv_c_and_k_pe_cache[1],
|
||||
attn_metadata.decode.block_table,
|
||||
attn_metadata.decode.seq_lens, self.num_heads, self.scale,
|
||||
self.num_kv_heads)
|
||||
else:
|
||||
# The MLA_PA path will be used as default path in the future, `_npu_paged_attention_mla` will
|
||||
# be removed after the torch_npu contains `torch_npu.atb.npu_multi_head_latent_attention` become
|
||||
# public available
|
||||
assert len(kv_c_and_k_pe_cache) > 1
|
||||
if envs_ascend.VLLM_ASCEND_MLA_PA:
|
||||
attn_output = torch_npu.atb.npu_multi_head_latent_attention(
|
||||
q_nope, q_pe, kv_c_and_k_pe_cache[0],
|
||||
kv_c_and_k_pe_cache[1], attn_metadata.decode.block_table,
|
||||
attn_metadata.decode.seq_lens, self.num_heads, self.scale,
|
||||
self.num_kv_heads)
|
||||
else:
|
||||
q = torch.cat([q_nope, q_pe], dim=-1)
|
||||
attn_output = torch.empty(
|
||||
[num_tokens, self.num_heads, self.kv_lora_rank],
|
||||
dtype=q.dtype,
|
||||
device=q.device)
|
||||
k_cache = torch.cat(
|
||||
[kv_c_and_k_pe_cache[0], kv_c_and_k_pe_cache[1]], dim=-1)
|
||||
torch_npu._npu_paged_attention_mla(
|
||||
query=q,
|
||||
key_cache=k_cache,
|
||||
num_kv_heads=self.num_kv_heads,
|
||||
num_heads=self.num_heads,
|
||||
scale_value=self.scale,
|
||||
block_table=attn_metadata.decode.
|
||||
block_table, # type:ignore
|
||||
context_lens=attn_metadata.decode.seq_lens, # type:ignore
|
||||
mla_vheadsize=self.kv_lora_rank,
|
||||
out=attn_output)
|
||||
q = torch.cat([q_nope, q_pe], dim=-1)
|
||||
attn_output = torch.empty(
|
||||
[num_tokens, self.num_heads, self.kv_lora_rank],
|
||||
dtype=q.dtype,
|
||||
device=q.device)
|
||||
k_cache = torch.cat(
|
||||
[kv_c_and_k_pe_cache[0], kv_c_and_k_pe_cache[1]], dim=-1)
|
||||
torch_npu._npu_paged_attention_mla(
|
||||
query=q,
|
||||
key_cache=k_cache,
|
||||
num_kv_heads=self.num_kv_heads,
|
||||
num_heads=self.num_heads,
|
||||
scale_value=self.scale,
|
||||
block_table=attn_metadata.decode.block_table, # type:ignore
|
||||
context_lens=attn_metadata.decode.seq_lens, # type:ignore
|
||||
mla_vheadsize=self.kv_lora_rank,
|
||||
out=attn_output)
|
||||
current_ms_metadata = get_multistream_comm_context()
|
||||
if current_ms_metadata is None:
|
||||
return self._v_up_proj_and_o_proj(attn_output,
|
||||
enable_multistream_mla)
|
||||
return self._v_up_proj_and_o_proj(attn_output)
|
||||
else:
|
||||
current_ms_metadata.before_comm_event.record()
|
||||
with torch.npu.stream(current_ms_metadata.comm_stream):
|
||||
@@ -1103,19 +805,14 @@ class AscendMLAImpl(MLAAttentionImpl):
|
||||
kv_cache: Tuple[torch.Tensor],
|
||||
attn_metadata: M,
|
||||
output: Optional[torch.Tensor] = None,
|
||||
enable_multistream_mla: bool = False,
|
||||
ckq: Optional[torch.Tensor] = None,
|
||||
) -> torch.Tensor:
|
||||
assert output is not None, "Output tensor must be provided."
|
||||
if attn_metadata is None:
|
||||
# Profiling run.
|
||||
return output
|
||||
self.running_in_graph = self.torchair_graph_enabled and attn_metadata.attn_state in [
|
||||
AscendAttentionState.DecodeOnly, AscendAttentionState.SpecDecoding
|
||||
]
|
||||
self.running_chunkprefilll_with_torchair = self.torchair_graph_enabled and attn_metadata.attn_state == AscendAttentionState.ChunkedPrefill
|
||||
num_actual_toks = attn_metadata.num_actual_tokens
|
||||
if k_pe is None and not self.running_in_graph:
|
||||
if k_pe is None:
|
||||
kv_c, k_pe = self.kv_a_proj_with_mqa(
|
||||
hidden_states_or_kv_c_normed)[0].split(
|
||||
[self.kv_lora_rank, self.qk_rope_head_dim], dim=-1)
|
||||
@@ -1128,134 +825,55 @@ class AscendMLAImpl(MLAAttentionImpl):
|
||||
has_decode = attn_metadata.num_decodes > 0
|
||||
has_prefill = attn_metadata.num_prefills > 0
|
||||
num_decode_tokens = attn_metadata.num_decode_tokens
|
||||
if not self.running_in_graph:
|
||||
# Inputs and outputs may be padded for CUDA graphs
|
||||
output_padded = output
|
||||
output = output[:num_actual_toks, ...]
|
||||
if not self.torchair_graph_enabled:
|
||||
kv_c_normed = kv_c_normed[:num_actual_toks, ...]
|
||||
prefill_k_c_normed = kv_c_normed[num_decode_tokens:]
|
||||
if not self.running_in_graph:
|
||||
hidden_states_or_q_c = hidden_states_or_q_c[:num_actual_toks, ...]
|
||||
prefill_hs_or_q_c = hidden_states_or_q_c[num_decode_tokens:]
|
||||
decode_hs_or_q_c = hidden_states_or_q_c[:num_decode_tokens]
|
||||
prefill_hs = hidden_states_or_kv_c_normed[num_decode_tokens:]
|
||||
# if not self.torchair_graph_enabled:
|
||||
k_pe = k_pe[:num_actual_toks, ...]
|
||||
k_pe = k_pe.unsqueeze(1)
|
||||
decode_k_pe = k_pe[:num_decode_tokens]
|
||||
prefill_k_pe = k_pe[num_decode_tokens:]
|
||||
else:
|
||||
decode_hs_or_q_c = hidden_states_or_q_c
|
||||
# Inputs and outputs may be padded for CUDA graphs
|
||||
output_padded = output
|
||||
output = output[:num_actual_toks, ...]
|
||||
kv_c_normed = kv_c_normed[:num_actual_toks, ...]
|
||||
prefill_k_c_normed = kv_c_normed[num_decode_tokens:]
|
||||
hidden_states_or_q_c = hidden_states_or_q_c[:num_actual_toks, ...]
|
||||
prefill_hs_or_q_c = hidden_states_or_q_c[num_decode_tokens:]
|
||||
decode_hs_or_q_c = hidden_states_or_q_c[:num_decode_tokens]
|
||||
k_pe = k_pe[:num_actual_toks, ...]
|
||||
k_pe = k_pe.unsqueeze(1)
|
||||
decode_k_pe = k_pe[:num_decode_tokens]
|
||||
prefill_k_pe = k_pe[num_decode_tokens:]
|
||||
if has_decode:
|
||||
decode_k_nope = None
|
||||
assert attn_metadata.decode is not None
|
||||
if self.running_in_graph or self.running_chunkprefilll_with_torchair:
|
||||
cos = attn_metadata.decode.cos
|
||||
sin = attn_metadata.decode.sin
|
||||
if self.running_chunkprefilll_with_torchair:
|
||||
decode_hs = (
|
||||
hidden_states_or_kv_c_normed[:num_decode_tokens])
|
||||
slots = attn_metadata.slot_mapping[:num_decode_tokens]
|
||||
decode_k_pe, decode_k_nope, decode_kv = self.exec_kv(
|
||||
decode_hs, cos, sin, kv_cache, slots)
|
||||
else:
|
||||
with npu_stream_switch("mla_secondary",
|
||||
0,
|
||||
enabled=enable_multistream_mla):
|
||||
npu_wait_tensor(hidden_states_or_kv_c_normed,
|
||||
ckq,
|
||||
enabled=enable_multistream_mla)
|
||||
decode_k_pe, decode_k_nope, decode_kv = self.exec_kv(
|
||||
hidden_states_or_kv_c_normed, cos, sin, kv_cache,
|
||||
attn_metadata.slot_mapping)
|
||||
# Without explicitly controlling the order, IndexByTensor operations
|
||||
# would be placed after `matmul W_KV_T` hindering the overlapping of
|
||||
# KvRmsNormRopeCache and SingleRope.
|
||||
npu_wait_tensor(decode_hs_or_q_c,
|
||||
cos,
|
||||
enabled=enable_multistream_mla)
|
||||
npu_wait_tensor(decode_hs_or_q_c,
|
||||
sin,
|
||||
enabled=enable_multistream_mla)
|
||||
npu_wait_tensor(decode_hs_or_q_c,
|
||||
decode_kv,
|
||||
enabled=enable_multistream_mla)
|
||||
|
||||
decode_ql_nope, decode_q_pe = \
|
||||
self._q_proj_and_k_up_proj(decode_hs_or_q_c)
|
||||
if self.running_in_graph:
|
||||
with npu_stream_switch("mla_secondary",
|
||||
0,
|
||||
enabled=enable_multistream_mla):
|
||||
npu_wait_tensor(decode_q_pe,
|
||||
decode_k_pe,
|
||||
enabled=enable_multistream_mla)
|
||||
decode_q_pe = self.rope_single(decode_q_pe, cos, sin)
|
||||
elif self.running_chunkprefilll_with_torchair:
|
||||
decode_q_pe = self.rope_single(decode_q_pe, cos, sin)
|
||||
else:
|
||||
decode_q_pe[...], decode_k_pe[...] = self.rotary_emb(
|
||||
attn_metadata.decode.input_positions,
|
||||
decode_q_pe.contiguous(),
|
||||
decode_k_pe,
|
||||
max_seq_len=attn_metadata.decode.max_seq_lens)
|
||||
decode_q_pe[...], decode_k_pe[...] = self.rotary_emb(
|
||||
attn_metadata.decode.input_positions,
|
||||
decode_q_pe.contiguous(),
|
||||
decode_k_pe,
|
||||
max_seq_len=attn_metadata.decode.max_seq_lens)
|
||||
if has_prefill:
|
||||
assert attn_metadata.prefill is not None
|
||||
prefill_q = self.q_proj(prefill_hs_or_q_c)[0]\
|
||||
.view(-1, self.num_heads, self.qk_head_dim)
|
||||
prefill_q_pe = prefill_q[..., self.qk_nope_head_dim:]
|
||||
prefill_q_nope = prefill_q[..., :self.qk_nope_head_dim]
|
||||
if self.torchair_graph_enabled:
|
||||
num_tokens = prefill_hs_or_q_c.shape[0]
|
||||
cos = attn_metadata.prefill.cos
|
||||
sin = attn_metadata.prefill.sin
|
||||
|
||||
prefill_q_pe = self.rope_single(prefill_q_pe, cos, sin)
|
||||
prefill_k_pe, prefill_k_nope = self.exec_kv_prefill(
|
||||
prefill_hs, cos, sin, kv_cache,
|
||||
attn_metadata.slot_mapping[num_decode_tokens:])
|
||||
|
||||
kv_c_normed = prefill_k_nope[:num_actual_toks, ...]
|
||||
prefill_k_c_normed = prefill_k_nope
|
||||
prefill_k_pe = prefill_k_pe.view(num_tokens, self.num_kv_heads,
|
||||
-1)
|
||||
prefill_q = torch.cat([prefill_q_nope, prefill_q_pe], dim=-1)
|
||||
else:
|
||||
prefill_q_pe[...], prefill_k_pe[...] = self.rotary_emb(
|
||||
attn_metadata.prefill.input_positions,
|
||||
prefill_q_pe.contiguous(),
|
||||
prefill_k_pe,
|
||||
max_seq_len=attn_metadata.prefill.max_seq_lens)
|
||||
prefill_q_pe[...], prefill_k_pe[...] = self.rotary_emb(
|
||||
attn_metadata.prefill.input_positions,
|
||||
prefill_q_pe.contiguous(),
|
||||
prefill_k_pe,
|
||||
max_seq_len=attn_metadata.prefill.max_seq_lens)
|
||||
|
||||
assert len(
|
||||
kv_cache
|
||||
) > 1, "the number of kv cache should be greater than 1, namely (nope_cache and rope_cache)"
|
||||
if self.torchair_graph_enabled:
|
||||
if kv_cache[0].numel() > 0 and has_prefill:
|
||||
slots = attn_metadata.slot_mapping
|
||||
# NOTE: Separate the kv cache in advance to avoid OOM or other issues
|
||||
torch_npu._npu_reshape_and_cache(
|
||||
key=kv_c_normed.view(num_tokens, self.num_kv_heads, -1),
|
||||
value=prefill_k_pe,
|
||||
key_cache=kv_cache[0],
|
||||
value_cache=kv_cache[1],
|
||||
slot_indices=slots[num_decode_tokens:])
|
||||
else:
|
||||
kv_c_normed = kv_c_normed.view(
|
||||
[num_actual_toks, self.num_kv_heads, -1])
|
||||
torch_npu._npu_reshape_and_cache(
|
||||
key=kv_c_normed,
|
||||
value=k_pe,
|
||||
key_cache=kv_cache[0],
|
||||
value_cache=kv_cache[1],
|
||||
slot_indices=attn_metadata.slot_mapping)
|
||||
if not self.running_in_graph:
|
||||
o_proj_input_shape = (num_actual_toks,
|
||||
self.num_heads * self.v_head_dim)
|
||||
o_proj_input = torch.empty(o_proj_input_shape,
|
||||
dtype=hidden_states_or_q_c.dtype,
|
||||
device=hidden_states_or_q_c.device)
|
||||
kv_c_normed = kv_c_normed.view(
|
||||
[num_actual_toks, self.num_kv_heads, -1])
|
||||
torch_npu._npu_reshape_and_cache(
|
||||
key=kv_c_normed,
|
||||
value=k_pe,
|
||||
key_cache=kv_cache[0],
|
||||
value_cache=kv_cache[1],
|
||||
slot_indices=attn_metadata.slot_mapping)
|
||||
o_proj_input_shape = (num_actual_toks,
|
||||
self.num_heads * self.v_head_dim)
|
||||
o_proj_input = torch.empty(o_proj_input_shape,
|
||||
dtype=hidden_states_or_q_c.dtype,
|
||||
device=hidden_states_or_q_c.device)
|
||||
if has_prefill:
|
||||
# FIX: aicore move should be also placed on the comm stream in dbo,
|
||||
# otherwise it may affect the accuracy
|
||||
@@ -1274,17 +892,9 @@ class AscendMLAImpl(MLAAttentionImpl):
|
||||
o_proj_input[num_decode_tokens:] = output_prefill
|
||||
|
||||
if has_decode:
|
||||
if self.running_in_graph:
|
||||
return self._forward_decode(decode_ql_nope, decode_q_pe,
|
||||
decode_k_nope, decode_k_pe,
|
||||
kv_cache, attn_metadata,
|
||||
enable_multistream_mla)
|
||||
else:
|
||||
output_decode = self._forward_decode(decode_ql_nope,
|
||||
decode_q_pe,
|
||||
decode_k_nope,
|
||||
decode_k_pe, kv_cache,
|
||||
attn_metadata)
|
||||
output_decode = self._forward_decode(decode_ql_nope, decode_q_pe,
|
||||
decode_k_nope, decode_k_pe,
|
||||
kv_cache, attn_metadata)
|
||||
current_ms_metadata = get_multistream_comm_context()
|
||||
if current_ms_metadata is not None:
|
||||
with torch.npu.stream(current_ms_metadata.comm_stream):
|
||||
@@ -1293,23 +903,13 @@ class AscendMLAImpl(MLAAttentionImpl):
|
||||
o_proj_input[:num_decode_tokens] = output_decode
|
||||
|
||||
current_ms_metadata = get_multistream_comm_context()
|
||||
MAX_O_PROJ_PREFETCH_SIZE = 16 * 1024 * 1024 # 16MB
|
||||
if current_ms_metadata is None:
|
||||
npu_prefetch(self.o_proj.weight,
|
||||
o_proj_input,
|
||||
max_size=MAX_O_PROJ_PREFETCH_SIZE,
|
||||
enabled=enable_multistream_mla)
|
||||
|
||||
output[...] = self.o_proj(
|
||||
o_proj_input,
|
||||
is_prefill=True,
|
||||
is_force_scatter=self.enable_shared_expert_dp)[0]
|
||||
else:
|
||||
with torch.npu.stream(current_ms_metadata.comm_stream):
|
||||
npu_prefetch(self.o_proj.weight,
|
||||
o_proj_input,
|
||||
max_size=MAX_O_PROJ_PREFETCH_SIZE,
|
||||
enabled=enable_multistream_mla)
|
||||
output[...] = self.o_proj(
|
||||
o_proj_input,
|
||||
is_prefill=True,
|
||||
|
||||
@@ -235,12 +235,18 @@ class NPUPlatform(Platform):
|
||||
raise ValueError("vLLM Ascend does not support V0 engine.")
|
||||
|
||||
use_torchair = get_ascend_config().torchair_graph_config.enabled
|
||||
if use_mla:
|
||||
return "vllm_ascend.attention.mla_v1.AscendMLABackend"
|
||||
elif use_torchair:
|
||||
return "vllm_ascend.torchair.torchair_attention.AscendAttentionTorchairBackend"
|
||||
else:
|
||||
return "vllm_ascend.attention.attention_v1.AscendAttentionBackend"
|
||||
# choose attention backend based on use_mla and use_torchair
|
||||
backend_map = {
|
||||
(True, True):
|
||||
"vllm_ascend.torchair.torchair_mla.AscendMLATorchairBackend",
|
||||
(True, False):
|
||||
"vllm_ascend.attention.mla_v1.AscendMLABackend",
|
||||
(False, True):
|
||||
"vllm_ascend.torchair.torchair_attention.AscendAttentionTorchairBackend",
|
||||
(False, False):
|
||||
"vllm_ascend.attention.attention_v1.AscendAttentionBackend"
|
||||
}
|
||||
return backend_map[(use_mla, use_torchair)]
|
||||
|
||||
@classmethod
|
||||
def get_punica_wrapper(cls) -> str:
|
||||
|
||||
1319
vllm_ascend/torchair/torchair_mla.py
Normal file
1319
vllm_ascend/torchair/torchair_mla.py
Normal file
File diff suppressed because it is too large
Load Diff
@@ -92,6 +92,7 @@ from vllm_ascend.multistream.ms_split import compute_split_seq_index
|
||||
from vllm_ascend.platform import NPUPlatform
|
||||
from vllm_ascend.sample.rejection_sampler import AscendRejectionSampler
|
||||
from vllm_ascend.torchair.torchair_attention import AscendTorchairMetadata
|
||||
from vllm_ascend.torchair.torchair_mla import AscendMLATorchairMetadata
|
||||
from vllm_ascend.utils import (ACL_FORMAT_FRACTAL_ND, ACL_FORMAT_FRACTAL_NZ,
|
||||
ProfileExecuteDuration, is_310p,
|
||||
maybe_converting_weight_acl_format)
|
||||
@@ -624,7 +625,7 @@ class NPUModelRunner(LoRAModelRunnerMixin):
|
||||
self,
|
||||
scheduler_output: "SchedulerOutput",
|
||||
) -> dict[str, Union[AscendMetadata, AscendMLAMetadata,
|
||||
AscendTorchairMetadata]]:
|
||||
AscendTorchairMetadata, AscendMLATorchairMetadata]]:
|
||||
total_num_scheduled_tokens = scheduler_output.total_num_scheduled_tokens
|
||||
assert total_num_scheduled_tokens > 0
|
||||
num_reqs = self.input_batch.num_reqs
|
||||
@@ -736,7 +737,8 @@ class NPUModelRunner(LoRAModelRunnerMixin):
|
||||
self.query_start_loc[num_reqs + 1:].fill_(-1)
|
||||
|
||||
attn_metadata: dict[str, Union[AscendMetadata, AscendMLAMetadata,
|
||||
AscendTorchairMetadata]] = {}
|
||||
AscendTorchairMetadata,
|
||||
AscendMLATorchairMetadata]] = {}
|
||||
# Prepare the attention metadata for each KV cache group and make layers
|
||||
# in the same group share the same metadata.
|
||||
for kv_cache_group_id, kv_cache_group_spec in enumerate(
|
||||
@@ -1000,8 +1002,8 @@ class NPUModelRunner(LoRAModelRunnerMixin):
|
||||
self,
|
||||
scheduler_output: "SchedulerOutput",
|
||||
intermediate_tensors: Optional[IntermediateTensors] = None,
|
||||
) -> tuple[Union[AscendMetadata, AscendMLAMetadata,
|
||||
AscendTorchairMetadata], torch.Tensor, np.ndarray, int,
|
||||
) -> tuple[Union[AscendMetadata, AscendMLAMetadata, AscendTorchairMetadata,
|
||||
AscendMLATorchairMetadata], torch.Tensor, np.ndarray, int,
|
||||
torch.Tensor, int, torch.Tensor, SpecDecodeMetadata,
|
||||
Optional[torch.Tensor], Optional[torch.Tensor],
|
||||
Optional[torch.Tensor]]:
|
||||
@@ -1466,7 +1468,8 @@ class NPUModelRunner(LoRAModelRunnerMixin):
|
||||
num_scheduled_tokens: int,
|
||||
hidden_states: torch.Tensor,
|
||||
attn_metadata: Union[AscendMetadata, AscendMLAMetadata,
|
||||
AscendTorchairMetadata],
|
||||
AscendTorchairMetadata,
|
||||
AscendMLATorchairMetadata],
|
||||
aux_hidden_states: torch.Tensor = None,
|
||||
) -> Optional[list[list[int]]]:
|
||||
if not self.use_spec_decode:
|
||||
@@ -2540,7 +2543,8 @@ class NPUModelRunner(LoRAModelRunnerMixin):
|
||||
num_scheduled_tokens: int,
|
||||
hidden_states: torch.Tensor,
|
||||
attn_metadata: Union[AscendMetadata, AscendMLAMetadata,
|
||||
AscendTorchairMetadata],
|
||||
AscendTorchairMetadata,
|
||||
AscendMLATorchairMetadata],
|
||||
):
|
||||
assert isinstance(self.drafter, MtpProposer)
|
||||
next_token_ids: list[int] = []
|
||||
|
||||
Reference in New Issue
Block a user