[Minor] move triton attention kernels into a separate folder (#1379)

This commit is contained in:
Lianmin Zheng
2024-09-10 15:15:08 -07:00
committed by GitHub
parent fbb4754cb8
commit 3a6e8b6d78
13 changed files with 24 additions and 15 deletions

View File

@@ -6,8 +6,11 @@ from flashinfer import (
)
from flashinfer.decode import _grouped_size_compiled_for_decode_kernels
from sglang.srt.layers.extend_attention import extend_attention_fwd, redundant_attention
from sglang.srt.layers.token_attention import token_attention_fwd
from sglang.srt.layers.triton_attention.extend_attention import (
extend_attention_fwd,
redundant_attention,
)
flashinfer_prefill_wrapper = None
flashinfer_decode_wrapper = None