From 9120e83d03249a97f6256bb694b414281076e2c3 Mon Sep 17 00:00:00 2001 From: Xinyuan Tong <115166877+JustinTong0323@users.noreply.github.com> Date: Tue, 15 Jul 2025 00:12:45 -0700 Subject: [PATCH] fix: remove redundant rotary embedding cache recomputation in MiniCPM (#8022) --- python/sglang/srt/models/minicpm.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/python/sglang/srt/models/minicpm.py b/python/sglang/srt/models/minicpm.py index 2df170f38..e7c94c85d 100644 --- a/python/sglang/srt/models/minicpm.py +++ b/python/sglang/srt/models/minicpm.py @@ -138,8 +138,6 @@ class MiniCPMAttention(nn.Module): base=rope_theta, rope_scaling=rope_scaling, ) - # set rope as fp32 instead of bf16 - self.rotary_emb.cos_sin_cache = self.rotary_emb._compute_cos_sin_cache() self.attn = RadixAttention( self.num_heads, self.head_dim,