Fix enable chunked prefill for Llama4 (#5575)

This commit is contained in:
tarinkk
2025-04-20 20:01:30 -04:00
committed by GitHub
parent 417b44eba8
commit 9a7e83e899

View File

@@ -73,8 +73,11 @@ class ModelConfig:
) )
if enable_multimodal is None: if enable_multimodal is None:
if self.hf_config.architectures == "Llama4ForConditionalGeneration": if self.hf_config.architectures[0] == "Llama4ForConditionalGeneration":
enable_multimodal = False enable_multimodal = False
logger.info(
"Multimodal is disabled for Llama4. To enable it, set --enable-llama4-multimodal."
)
else: else:
enable_multimodal = True enable_multimodal = True