Update transformers package version to 4.57.0 (#11222)
Co-authored-by: yhyang201 <yhyang201@gmail.com>
This commit is contained in:
@@ -63,7 +63,7 @@ dependencies = [
|
||||
"torchaudio==2.8.0",
|
||||
"torchvision",
|
||||
"tqdm",
|
||||
"transformers==4.56.1",
|
||||
"transformers==4.57.0",
|
||||
"uvicorn",
|
||||
"uvloop",
|
||||
"xgrammar==0.1.24",
|
||||
|
||||
@@ -49,7 +49,7 @@ from typing import List, Optional, Sequence, Tuple, Union
|
||||
import torch
|
||||
import torch.nn as nn
|
||||
import torch.nn.functional as F
|
||||
from transformers.activations import ACT2FN, PytorchGELUTanh
|
||||
from transformers.activations import ACT2FN, GELUTanh
|
||||
from transformers.modeling_utils import PreTrainedModel
|
||||
|
||||
try:
|
||||
@@ -614,7 +614,7 @@ class MoonVitPretrainedModel(PreTrainedModel):
|
||||
"num_heads": config.num_attention_heads,
|
||||
"hidden_dim": config.hidden_size,
|
||||
"mlp_dim": config.intermediate_size,
|
||||
"activation": PytorchGELUTanh(),
|
||||
"activation": GELUTanh(),
|
||||
"attn_bias": True,
|
||||
"attn_implementation": config._attn_implementation,
|
||||
},
|
||||
|
||||
Reference in New Issue
Block a user