[Fix] Fix several issues preventing gemma3n LoRA support. (#8776)
This commit is contained in:
@@ -1943,10 +1943,16 @@ class ServerArgs:
|
||||
if "Llama4" in model_arch:
|
||||
assert self.attention_backend == "fa3", "fa3 is required for Llama4 model"
|
||||
|
||||
if "Gemma2ForCausalLM" in model_arch:
|
||||
if model_arch in [
|
||||
"Gemma2ForCausalLM",
|
||||
"Gemma3nForCausalLM",
|
||||
"Gemma3nForConditionalGeneration",
|
||||
]:
|
||||
# FIXME: https://github.com/sgl-project/sglang/pull/7367 is not compatible with gemma2 model.
|
||||
# It failed at this test: https://github.com/sgl-project/sglang/actions/runs/16255155597/job/45890331952#step:4:736
|
||||
logger.warning("Disable hybrid SWA memory for Gemma2ForCausalLM.")
|
||||
logger.warning(
|
||||
f"Disable hybrid SWA memory for {model_arch} as it is not yet supported."
|
||||
)
|
||||
self.disable_hybrid_swa_memory = True
|
||||
|
||||
# Check LoRA
|
||||
|
||||
Reference in New Issue
Block a user