From 9e0dac1ad706aa8769860796e60a218d6c20868b Mon Sep 17 00:00:00 2001 From: Liangsheng Yin Date: Thu, 17 Oct 2024 18:33:21 -0700 Subject: [PATCH] Fix regex and logprob conflicts when chunked prefilling (#1703) --- python/sglang/srt/managers/scheduler.py | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/python/sglang/srt/managers/scheduler.py b/python/sglang/srt/managers/scheduler.py index 9f6989c25..7ba163959 100644 --- a/python/sglang/srt/managers/scheduler.py +++ b/python/sglang/srt/managers/scheduler.py @@ -789,15 +789,15 @@ class Scheduler: elif not batch.decoding_reqs or req not in batch.decoding_reqs: self.tree_cache.cache_unfinished_req(req) - if req.regex_fsm is not None: - req.regex_fsm_state = req.regex_fsm.get_next_state( - req.regex_fsm_state, next_token_ids[i] - ) + if req.regex_fsm is not None: + req.regex_fsm_state = req.regex_fsm.get_next_state( + req.regex_fsm_state, next_token_ids[i] + ) - if req.return_logprob: - logprob_pt += self.add_logprob_return_values( - i, req, logprob_pt, next_token_ids, logits_output - ) + if req.return_logprob: + logprob_pt += self.add_logprob_return_values( + i, req, logprob_pt, next_token_ids, logits_output + ) else: # embedding or reward model assert batch.extend_num_tokens != 0 embeddings, bid = result