### What this PR does / why we need it?
Fixes `transformers_utils/processors/__init__` import error, due to
https://github.com/vllm-project/vllm/pull/33247
Fixes Fused MoE break introduced by `MoERunner abstraction,` due to
https://github.com/vllm-project/vllm/pull/32344
> delete AscendMoERunnere when
https://github.com/vllm-project/vllm/pull/35178 is merged
Fixes `Make Qwen3VL compatible with Transformers v5`, due to
https://github.com/vllm-project/vllm/pull/34262
### Does this PR introduce _any_ user-facing change?
### How was this patch tested?
- vLLM version: v0.15.0
- vLLM main:
9562912cea
---------
Signed-off-by: wxsIcey <1790571317@qq.com>
39 lines
645 B
Plaintext
39 lines
645 B
Plaintext
# Should be mirrored in pyporject.toml
|
|
cmake>=3.26
|
|
decorator
|
|
einops
|
|
numpy<2.0.0
|
|
packaging
|
|
pip
|
|
pybind11
|
|
pyyaml
|
|
scipy
|
|
pandas
|
|
setuptools>=64
|
|
setuptools-scm>=8
|
|
torch==2.9.0
|
|
torchvision
|
|
torchaudio
|
|
wheel
|
|
xgrammar>=0.1.30
|
|
pandas-stubs
|
|
opencv-python-headless<=4.11.0.86 # Required to avoid numpy version conflict with vllm
|
|
compressed_tensors>=0.11.0
|
|
|
|
# requirements for disaggregated prefill
|
|
msgpack
|
|
quart
|
|
|
|
# Required for N-gram speculative decoding
|
|
numba
|
|
|
|
# Install torch_npu
|
|
#--pre
|
|
#--extra-index-url https://mirrors.huaweicloud.com/ascend/repos/pypi
|
|
torch-npu==2.9.0
|
|
|
|
arctic-inference==0.1.1
|
|
transformers>=4.57.4
|
|
fastapi<0.124.0
|
|
triton-ascend==3.2.0
|