From de2faef97e610c35211847b24f7512860d80a2d8 Mon Sep 17 00:00:00 2001 From: Ke Bao Date: Fri, 2 May 2025 00:28:46 +0800 Subject: [PATCH] Remove extra contiguous (#5953) --- python/sglang/srt/models/deepseek_v2.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/python/sglang/srt/models/deepseek_v2.py b/python/sglang/srt/models/deepseek_v2.py index 5ce693efa..85695b114 100644 --- a/python/sglang/srt/models/deepseek_v2.py +++ b/python/sglang/srt/models/deepseek_v2.py @@ -752,7 +752,7 @@ class DeepseekV2AttentionMLA(nn.Module): q_nope_out = q_nope_out.transpose(0, 1) k_nope = latent_cache[..., : self.kv_lora_rank] - k_nope = self.kv_a_layernorm(k_nope.contiguous()).unsqueeze(1) + k_nope = self.kv_a_layernorm(k_nope).unsqueeze(1) k_pe = latent_cache[..., self.kv_lora_rank :].unsqueeze(1) q_pe, k_pe = self.rotary_emb(positions, q_pe, k_pe)