[bugfix] Add modelscope package to avoid docker image without modelscope (#1520)
This commit is contained in:
11
README.md
11
README.md
@@ -287,6 +287,17 @@ Launch [Qwen2-7B-Instruct](https://www.modelscope.cn/models/qwen/qwen2-7b-instru
|
|||||||
SGLANG_USE_MODELSCOPE=true python -m sglang.launch_server --model-path qwen/Qwen2-7B-Instruct --port 30000
|
SGLANG_USE_MODELSCOPE=true python -m sglang.launch_server --model-path qwen/Qwen2-7B-Instruct --port 30000
|
||||||
```
|
```
|
||||||
|
|
||||||
|
Or start it by docker.
|
||||||
|
```bash
|
||||||
|
docker run --gpus all \
|
||||||
|
-p 30000:30000 \
|
||||||
|
-v ~/.cache/modelscope:/root/.cache/modelscope \
|
||||||
|
--env "SGLANG_USE_MODELSCOPE=true" \
|
||||||
|
--ipc=host \
|
||||||
|
lmsysorg/sglang:latest \
|
||||||
|
python3 -m sglang.launch_server --model-path Qwen/Qwen2.5-7B-Instruct --host 0.0.0.0 --port 30000
|
||||||
|
```
|
||||||
|
|
||||||
</details>
|
</details>
|
||||||
|
|
||||||
#### Run Llama 3.1 405B
|
#### Run Llama 3.1 405B
|
||||||
|
|||||||
@@ -4,6 +4,8 @@ services:
|
|||||||
container_name: sglang
|
container_name: sglang
|
||||||
volumes:
|
volumes:
|
||||||
- ${HOME}/.cache/huggingface:/root/.cache/huggingface
|
- ${HOME}/.cache/huggingface:/root/.cache/huggingface
|
||||||
|
# If you use modelscope, you need mount this directory
|
||||||
|
# - ${HOME}/.cache/modelscope:/root/.cache/modelscope
|
||||||
restart: always
|
restart: always
|
||||||
network_mode: host
|
network_mode: host
|
||||||
# Or you can only publish port 30000
|
# Or you can only publish port 30000
|
||||||
@@ -11,6 +13,8 @@ services:
|
|||||||
# - 30000:30000
|
# - 30000:30000
|
||||||
environment:
|
environment:
|
||||||
HF_TOKEN: <secret>
|
HF_TOKEN: <secret>
|
||||||
|
# if you use modelscope to download model, you need set this environment
|
||||||
|
# - SGLANG_USE_MODELSCOPE: true
|
||||||
entrypoint: python3 -m sglang.launch_server
|
entrypoint: python3 -m sglang.launch_server
|
||||||
command:
|
command:
|
||||||
--model-path meta-llama/Meta-Llama-3.1-8B-Instruct
|
--model-path meta-llama/Meta-Llama-3.1-8B-Instruct
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ dependencies = [
|
|||||||
srt = ["aiohttp", "decord", "fastapi", "hf_transfer", "huggingface_hub", "interegular",
|
srt = ["aiohttp", "decord", "fastapi", "hf_transfer", "huggingface_hub", "interegular",
|
||||||
"packaging", "pillow", "psutil", "pydantic", "python-multipart",
|
"packaging", "pillow", "psutil", "pydantic", "python-multipart",
|
||||||
"torch", "torchao", "uvicorn", "uvloop", "zmq",
|
"torch", "torchao", "uvicorn", "uvloop", "zmq",
|
||||||
"vllm==0.5.5", "outlines>=0.0.44"]
|
"vllm==0.5.5", "outlines>=0.0.44", "modelscope"]
|
||||||
openai = ["openai>=1.0", "tiktoken"]
|
openai = ["openai>=1.0", "tiktoken"]
|
||||||
anthropic = ["anthropic>=0.20.0"]
|
anthropic = ["anthropic>=0.20.0"]
|
||||||
litellm = ["litellm>=1.0.0"]
|
litellm = ["litellm>=1.0.0"]
|
||||||
|
|||||||
Reference in New Issue
Block a user