[Bug]: Add missing clamp to llavavid (#3787)

This commit is contained in:
Yueyang Pan
2025-02-25 04:10:15 +01:00
committed by GitHub
parent 6ce9dbe828
commit 7036d6fc67

View File

@@ -110,6 +110,9 @@ class LlavaVidForCausalLM(nn.Module):
if forward_batch.forward_mode.is_extend():
bs = forward_batch.batch_size
# Clamp input ids. See llava.py for more details
input_ids = input_ids.clamp_(min=0, max=self.config.vocab_size - 1)
# Embed text inputs
input_embeds = self.language_model.model.embed_tokens(input_ids)