diff --git a/python/sglang/srt/layers/attention/flashmla_backend.py b/python/sglang/srt/layers/attention/flashmla_backend.py index 1513c1c71..0823239a7 100644 --- a/python/sglang/srt/layers/attention/flashmla_backend.py +++ b/python/sglang/srt/layers/attention/flashmla_backend.py @@ -241,6 +241,9 @@ class FlashMLABackend(FlashInferMLAAttnBackend): seq_lens_cpu, ) + def get_cuda_graph_seq_len_fill_value(self): + return 1024 + def forward_decode( self, q: torch.Tensor,