Fix srt dependency (#1685)
This commit is contained in:
@@ -24,11 +24,9 @@ runtime_common = ["aiohttp", "decord", "fastapi", "hf_transfer", "huggingface_hu
|
|||||||
"packaging", "pillow", "psutil", "pydantic", "python-multipart",
|
"packaging", "pillow", "psutil", "pydantic", "python-multipart",
|
||||||
"torchao", "uvicorn", "uvloop", "zmq",
|
"torchao", "uvicorn", "uvloop", "zmq",
|
||||||
"outlines>=0.0.44", "modelscope"]
|
"outlines>=0.0.44", "modelscope"]
|
||||||
torch = ["torch"]
|
|
||||||
# xpu is not enabled in public vllm and torch whl,
|
# xpu is not enabled in public vllm and torch whl,
|
||||||
# need to follow https://docs.vllm.ai/en/latest/getting_started/xpu-installation.htmlinstall vllm
|
# need to follow https://docs.vllm.ai/en/latest/getting_started/xpu-installation.htmlinstall vllm
|
||||||
vllm = ["vllm==0.5.5"]
|
srt = ["sglang[runtime_common]", "torch", "vllm==0.5.5"]
|
||||||
srt = ["sglang[runtime_common]", "torch", "vllm"]
|
|
||||||
srt_xpu = ["sglang[runtime_common]"]
|
srt_xpu = ["sglang[runtime_common]"]
|
||||||
|
|
||||||
openai = ["openai>=1.0", "tiktoken"]
|
openai = ["openai>=1.0", "tiktoken"]
|
||||||
|
|||||||
Reference in New Issue
Block a user