From 9a7e83e899112b5236b7d6ae2d7b202457b51902 Mon Sep 17 00:00:00 2001 From: tarinkk <129432511+tarinkk@users.noreply.github.com> Date: Sun, 20 Apr 2025 20:01:30 -0400 Subject: [PATCH] Fix enable chunked prefill for Llama4 (#5575) --- python/sglang/srt/configs/model_config.py | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/python/sglang/srt/configs/model_config.py b/python/sglang/srt/configs/model_config.py index 7aaee1547..28bf9c83e 100644 --- a/python/sglang/srt/configs/model_config.py +++ b/python/sglang/srt/configs/model_config.py @@ -73,8 +73,11 @@ class ModelConfig: ) if enable_multimodal is None: - if self.hf_config.architectures == "Llama4ForConditionalGeneration": + if self.hf_config.architectures[0] == "Llama4ForConditionalGeneration": enable_multimodal = False + logger.info( + "Multimodal is disabled for Llama4. To enable it, set --enable-llama4-multimodal." + ) else: enable_multimodal = True