From 77a6c9d22927fa425e89cdcf672cc61c14b72eaf Mon Sep 17 00:00:00 2001 From: Chaitanya Sri Krishna Lolla Date: Fri, 28 Feb 2025 23:31:08 +0530 Subject: [PATCH] Remove unused imports from rocm mla kernel. (#3963) --- .../layers/attention/triton_ops/rocm_mla_decode_rope.py | 7 ------- 1 file changed, 7 deletions(-) diff --git a/python/sglang/srt/layers/attention/triton_ops/rocm_mla_decode_rope.py b/python/sglang/srt/layers/attention/triton_ops/rocm_mla_decode_rope.py index 4c3e63968..915f4ef92 100644 --- a/python/sglang/srt/layers/attention/triton_ops/rocm_mla_decode_rope.py +++ b/python/sglang/srt/layers/attention/triton_ops/rocm_mla_decode_rope.py @@ -20,19 +20,12 @@ It supports page size = 1. # https://github.com/ModelTC/lightllm/blob/96353e868a840db4d103138caf15ed9dbea8c186/lightllm/models/deepseek2/triton_kernel/gqa_flash_decoding_stage1.py # https://github.com/ModelTC/lightllm/blob/96353e868a840db4d103138caf15ed9dbea8c186/lightllm/models/deepseek2/triton_kernel/gqa_flash_decoding_stage2.py -import argparse -import logging -import sys - -import pytest -import torch import triton import triton.language as tl from sglang.srt.layers.attention.triton_ops.decode_attention import ( _decode_softmax_reducev_fwd, ) -from sglang.srt.layers.rotary_embedding import DeepseekScalingRotaryEmbedding def is_hip():