From 556e4143f0ba8e480a73fc0e08f05d7ea83880cc Mon Sep 17 00:00:00 2001 From: Yineng Zhang Date: Tue, 5 Aug 2025 13:40:22 -0700 Subject: [PATCH] fix: remove unused import (#8809) --- python/sglang/srt/layers/attention/trtllm_mha_backend.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/python/sglang/srt/layers/attention/trtllm_mha_backend.py b/python/sglang/srt/layers/attention/trtllm_mha_backend.py index 2e7c67758..1d197c5da 100644 --- a/python/sglang/srt/layers/attention/trtllm_mha_backend.py +++ b/python/sglang/srt/layers/attention/trtllm_mha_backend.py @@ -1,7 +1,5 @@ from __future__ import annotations -from python.sglang.srt.layers.radix_attention import RadixAttention - """ Support attention backend for TRTLLM MLA kernels from flashinfer. """