diff --git a/python/sglang/srt/layers/attention/trtllm_mha_backend.py b/python/sglang/srt/layers/attention/trtllm_mha_backend.py index 2e7c67758..1d197c5da 100644 --- a/python/sglang/srt/layers/attention/trtllm_mha_backend.py +++ b/python/sglang/srt/layers/attention/trtllm_mha_backend.py @@ -1,7 +1,5 @@ from __future__ import annotations -from python.sglang.srt.layers.radix_attention import RadixAttention - """ Support attention backend for TRTLLM MLA kernels from flashinfer. """