Fix enable chunked prefill for Llama4 (#5575)
This commit is contained in:
@@ -73,8 +73,11 @@ class ModelConfig:
|
|||||||
)
|
)
|
||||||
|
|
||||||
if enable_multimodal is None:
|
if enable_multimodal is None:
|
||||||
if self.hf_config.architectures == "Llama4ForConditionalGeneration":
|
if self.hf_config.architectures[0] == "Llama4ForConditionalGeneration":
|
||||||
enable_multimodal = False
|
enable_multimodal = False
|
||||||
|
logger.info(
|
||||||
|
"Multimodal is disabled for Llama4. To enable it, set --enable-llama4-multimodal."
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
enable_multimodal = True
|
enable_multimodal = True
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user