[Misc] Upgrade vllm hash to 12_14 (#5000)
### What this PR does / why we need it?
### Does this PR introduce _any_ user-facing change?
1. fix https://github.com/vllm-project/vllm/pull/27938
2. fix https://github.com/vllm-project/vllm/pull/27145
pooling models now supports chunked prefill and prefix caching,
3. fix https://github.com/vllm-project/vllm/pull/30181
define the CPU fields in the field config where they really belong.
4. fix https://github.com/vllm-project/vllm/pull/28168
define the CPU fields in the field config where they really belong.
5. fix https://github.com/vllm-project/vllm/pull/30201
some moudle rename
6. fix https://github.com/vllm-project/vllm/pull/29067
fusedmoe moudle refactor
7. fix https://github.com/vllm-project/vllm/pull/29066
fusedmoe moudle refactor
8. fix https://github.com/vllm-project/vllm/pull/29624
### How was this patch tested?
- vLLM version: v0.12.0
- vLLM main:
ad32e3e19c
---------
Signed-off-by: wangli <wangli858794774@gmail.com>
This commit is contained in:
@@ -365,6 +365,10 @@ class NPUPlatform(Platform):
|
||||
use_mla,
|
||||
has_sink=False,
|
||||
use_sparse=False,
|
||||
# NOTE: Please pay special attention to the order of these parameters.
|
||||
# Although we are only using some of them so far
|
||||
# vllm passes them in sequence when using this interface.
|
||||
use_mm_prefix: bool = False,
|
||||
attn_type: str | None = None,
|
||||
):
|
||||
# choose attention backend based on use_mla
|
||||
|
||||
Reference in New Issue
Block a user