From 035ac2ab74229d4963abd7768e12d2d8bd729402 Mon Sep 17 00:00:00 2001 From: Mick Date: Sun, 16 Mar 2025 04:27:26 +0800 Subject: [PATCH] ci: update transformers==4.48.3 (#4451) --- .../srt/managers/image_processors/base_image_processor.py | 7 ------- scripts/ci_install_dependency.sh | 2 +- 2 files changed, 1 insertion(+), 8 deletions(-) diff --git a/python/sglang/srt/managers/image_processors/base_image_processor.py b/python/sglang/srt/managers/image_processors/base_image_processor.py index ec1799b69..c4349d16c 100644 --- a/python/sglang/srt/managers/image_processors/base_image_processor.py +++ b/python/sglang/srt/managers/image_processors/base_image_processor.py @@ -23,13 +23,6 @@ def get_global_processor(): return global_processor -def init_global_processor(sglang_image_processor, server_args: ServerArgs): - """Init the global processor for multi-modal models.""" - global global_processor - transformers.logging.set_verbosity_error() - global_processor = sglang_image_processor._build_processor(server_args=server_args) - - @dataclasses.dataclass class BaseImageProcessorOutput: image_hashes: list[int] diff --git a/scripts/ci_install_dependency.sh b/scripts/ci_install_dependency.sh index 3bf8050db..b1d14973e 100755 --- a/scripts/ci_install_dependency.sh +++ b/scripts/ci_install_dependency.sh @@ -20,7 +20,7 @@ pip install flashinfer_python==0.2.3 --find-links ${FLASHINFER_REPO} --force-rei pip install torch_memory_saver --force-reinstall -pip install transformers==4.45.2 sentence_transformers accelerate==1.4.0 peft pandas datasets +pip install transformers==4.48.3 sentence_transformers accelerate==1.4.0 peft pandas datasets # For compling xgrammar kernels pip install cuda-python nvidia-cuda-nvrtc-cu12