200 lines
6.4 KiB
Python
200 lines
6.4 KiB
Python
import logging
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
ATTENTION_BACKENDS = {}
|
|
|
|
|
|
def register_attention_backend(name):
|
|
def decorator(fn):
|
|
ATTENTION_BACKENDS[name] = fn
|
|
return fn
|
|
|
|
return decorator
|
|
|
|
|
|
@register_attention_backend("flashinfer")
|
|
def create_flashinfer_backend(runner):
|
|
import torch
|
|
|
|
if not runner.use_mla_backend:
|
|
from sglang.srt.layers.attention.flashinfer_backend import FlashInferAttnBackend
|
|
|
|
# Init streams
|
|
if runner.server_args.speculative_algorithm == "EAGLE":
|
|
if (
|
|
not hasattr(runner, "plan_stream_for_flashinfer")
|
|
or not runner.plan_stream_for_flashinfer
|
|
):
|
|
runner.plan_stream_for_flashinfer = torch.cuda.Stream()
|
|
return FlashInferAttnBackend(runner)
|
|
else:
|
|
from sglang.srt.layers.attention.flashinfer_mla_backend import (
|
|
FlashInferMLAAttnBackend,
|
|
)
|
|
|
|
return FlashInferMLAAttnBackend(runner)
|
|
|
|
|
|
@register_attention_backend("trtllm_mla")
|
|
def create_trtllm_mla_backend(runner):
|
|
if not runner.use_mla_backend:
|
|
raise ValueError("trtllm_mla backend can only be used with MLA models.")
|
|
from sglang.srt.layers.attention.trtllm_mla_backend import TRTLLMMLABackend
|
|
|
|
return TRTLLMMLABackend(runner)
|
|
|
|
|
|
@register_attention_backend("aiter")
|
|
def create_aiter_backend(runner):
|
|
from sglang.srt.layers.attention.aiter_backend import AiterAttnBackend
|
|
|
|
return AiterAttnBackend(runner)
|
|
|
|
|
|
@register_attention_backend("wave")
|
|
def create_wave_backend(runner):
|
|
from sglang.srt.layers.attention.wave_backend import WaveAttnBackend
|
|
|
|
return WaveAttnBackend(runner)
|
|
|
|
|
|
@register_attention_backend("ascend")
|
|
def create_ascend_backend(runner):
|
|
from sglang.srt.layers.attention.ascend_backend import AscendAttnBackend
|
|
|
|
return AscendAttnBackend(runner)
|
|
|
|
|
|
@register_attention_backend("triton")
|
|
def create_triton_backend(runner):
|
|
assert not runner.model_config.is_encoder_decoder, (
|
|
"Cross attention is not supported in the triton attention backend. "
|
|
"Please use `--attention-backend flashinfer`."
|
|
)
|
|
if runner.server_args.enable_double_sparsity:
|
|
from sglang.srt.layers.attention.double_sparsity_backend import (
|
|
DoubleSparseAttnBackend,
|
|
)
|
|
|
|
return DoubleSparseAttnBackend(runner)
|
|
else:
|
|
from sglang.srt.layers.attention.triton_backend import TritonAttnBackend
|
|
|
|
return TritonAttnBackend(runner)
|
|
|
|
|
|
@register_attention_backend("torch_native")
|
|
def create_torch_native_backend(runner):
|
|
from sglang.srt.layers.attention.torch_native_backend import TorchNativeAttnBackend
|
|
|
|
return TorchNativeAttnBackend(runner)
|
|
|
|
|
|
@register_attention_backend("flex_attention")
|
|
def create_flex_attention_backend(runner):
|
|
from sglang.srt.layers.attention.torch_flex_backend import TorchFlexAttnBackend
|
|
|
|
return TorchFlexAttnBackend(runner)
|
|
|
|
|
|
@register_attention_backend("flashmla")
|
|
def create_flashmla_backend(runner):
|
|
from sglang.srt.layers.attention.flashmla_backend import FlashMLABackend
|
|
|
|
return FlashMLABackend(runner)
|
|
|
|
|
|
@register_attention_backend("fa3")
|
|
def create_flashattention_v3_backend(runner):
|
|
import torch
|
|
|
|
assert (
|
|
torch.cuda.get_device_capability()[0] == 8 and not runner.use_mla_backend
|
|
) or torch.cuda.get_device_capability()[0] == 9, (
|
|
"FlashAttention v3 Backend requires SM>=80 and SM<=90. "
|
|
"Please use `--attention-backend flashinfer`."
|
|
)
|
|
from sglang.srt.layers.attention.flashattention_backend import FlashAttentionBackend
|
|
|
|
return FlashAttentionBackend(runner)
|
|
|
|
|
|
@register_attention_backend("fa4")
|
|
def create_flashattention_v4_backend(runner):
|
|
assert (
|
|
runner.use_mla_backend
|
|
), "FlashAttention v4 Support is at an early stage, only MLA model supported now"
|
|
from sglang.srt.layers.attention.flashattention_backend import FlashAttentionBackend
|
|
|
|
return FlashAttentionBackend(runner, fa_impl_ver=4)
|
|
|
|
|
|
@register_attention_backend("cutlass_mla")
|
|
def create_cutlass_mla_backend(runner):
|
|
from sglang.srt.layers.attention.cutlass_mla_backend import CutlassMLABackend
|
|
|
|
return CutlassMLABackend(runner)
|
|
|
|
|
|
@register_attention_backend("trtllm_mha")
|
|
def create_trtllm_mha_backend(runner):
|
|
if runner.use_mla_backend:
|
|
raise ValueError("trtllm_mha backend can only be used with non-MLA models.")
|
|
from sglang.srt.layers.attention.trtllm_mha_backend import TRTLLMHAAttnBackend
|
|
|
|
return TRTLLMHAAttnBackend(runner)
|
|
|
|
|
|
@register_attention_backend("intel_amx")
|
|
def create_intel_amx_backend(runner):
|
|
from sglang.srt.layers.attention.intel_amx_backend import IntelAMXAttnBackend
|
|
|
|
return IntelAMXAttnBackend(runner)
|
|
|
|
|
|
@register_attention_backend("dual_chunk_flash_attn")
|
|
def create_dual_chunk_flash_attn_backend(runner):
|
|
from sglang.srt.layers.attention.dual_chunk_flashattention_backend import (
|
|
DualChunkFlashAttentionBackend,
|
|
)
|
|
|
|
return DualChunkFlashAttentionBackend(runner)
|
|
|
|
|
|
def attn_backend_wrapper(runner, full_attn_backend):
|
|
"""
|
|
Wrapper for special models like hybrid GDN, so we don't
|
|
need to change the code of the original attention backend.
|
|
"""
|
|
assert not (
|
|
runner.is_hybrid_gdn and runner.use_mla_backend
|
|
), "hybrid_gdn can only be used with non-MLA models."
|
|
|
|
# wrap for hybrid GDN models
|
|
if runner.is_hybrid_gdn:
|
|
from sglang.srt.utils import is_blackwell, is_npu
|
|
|
|
if is_blackwell():
|
|
assert (
|
|
runner.server_args.attention_backend == "triton"
|
|
or runner.server_args.attention_backend == "trtllm_mha"
|
|
), "triton or trtllm_mha backend are the only supported backends on Blackwell GPUs for hybrid GDN models, use --attention-backend triton or --attention-backend trtllm_mha to specify the backend."
|
|
if is_npu():
|
|
assert (
|
|
runner.server_args.attention_backend == "ascend"
|
|
), "ascend backend is the only supported backend on NPU for hybrid GDN models, use --attention-backend ascend to specify the backend."
|
|
logger.info(f"Using hybrid linear attention backend for hybrid GDN models.")
|
|
from sglang.srt.layers.attention.hybrid_linear_attn_backend import (
|
|
HybridLinearAttnBackend,
|
|
MambaAttnBackend,
|
|
)
|
|
|
|
linear_attn_backend = MambaAttnBackend(runner)
|
|
full_attn_layers = runner.model_config.hf_config.full_attention_layer_ids
|
|
return HybridLinearAttnBackend(
|
|
full_attn_backend, linear_attn_backend, full_attn_layers
|
|
)
|
|
|
|
return full_attn_backend
|