[Fix] Fix several issues preventing gemma3n LoRA support. (#8776)

This commit is contained in:
Lifu Huang
2025-08-04 17:11:46 -07:00
committed by GitHub
parent 6d0646da11
commit 7cb20754fa
3 changed files with 54 additions and 2 deletions

View File

@@ -1943,10 +1943,16 @@ class ServerArgs:
if "Llama4" in model_arch:
assert self.attention_backend == "fa3", "fa3 is required for Llama4 model"
if "Gemma2ForCausalLM" in model_arch:
if model_arch in [
"Gemma2ForCausalLM",
"Gemma3nForCausalLM",
"Gemma3nForConditionalGeneration",
]:
# FIXME: https://github.com/sgl-project/sglang/pull/7367 is not compatible with gemma2 model.
# It failed at this test: https://github.com/sgl-project/sglang/actions/runs/16255155597/job/45890331952#step:4:736
logger.warning("Disable hybrid SWA memory for Gemma2ForCausalLM.")
logger.warning(
f"Disable hybrid SWA memory for {model_arch} as it is not yet supported."
)
self.disable_hybrid_swa_memory = True
# Check LoRA