[Doc] Update tutorials for single_npu_audio and single_npu_multimodal (#2252)
### What this PR does / why we need it?
Update tutorials for single_npu_audio and single_npu_multimodal
- vLLM version: v0.10.0
- vLLM main:
6b47ef24de
Signed-off-by: hfadzxy <starmoon_zhang@163.com>
This commit is contained in:
@@ -54,6 +54,7 @@ from vllm import LLM, SamplingParams
|
||||
from vllm.assets.audio import AudioAsset
|
||||
from vllm.utils import FlexibleArgumentParser
|
||||
|
||||
# If network issues prevent AudioAsset from fetching remote audio files, retry or check your network.
|
||||
audio_assets = [AudioAsset("mary_had_lamb"), AudioAsset("winning_call")]
|
||||
question_per_audio_count = {
|
||||
1: "What is recited in the audio?",
|
||||
|
||||
@@ -43,7 +43,7 @@ export PYTORCH_NPU_ALLOC_CONF=max_split_size_mb:256
|
||||
Run the following script to execute offline inference on a single NPU:
|
||||
|
||||
```bash
|
||||
pip install torchvision==0.20.1 qwen_vl_utils --extra-index-url https://download.pytorch.org/whl/cpu/
|
||||
pip install qwen_vl_utils --extra-index-url https://download.pytorch.org/whl/cpu/
|
||||
```
|
||||
|
||||
```python
|
||||
|
||||
Reference in New Issue
Block a user