forked from EngineX-Cambricon/enginex-mlu370-vllm
add qwen3
This commit is contained in:
12
vllm-v0.6.2/docs/source/serving/deploying_with_lws.rst
Normal file
12
vllm-v0.6.2/docs/source/serving/deploying_with_lws.rst
Normal file
@@ -0,0 +1,12 @@
|
||||
.. _deploying_with_lws:
|
||||
|
||||
Deploying with LWS
|
||||
============================
|
||||
|
||||
LeaderWorkerSet (LWS) is a Kubernetes API that aims to address common deployment patterns of AI/ML inference workloads.
|
||||
A major use case is for multi-host/multi-node distributed inference.
|
||||
|
||||
vLLM can be deployed with `LWS <https://github.com/kubernetes-sigs/lws>`_ on Kubernetes for distributed model serving.
|
||||
|
||||
Please see `this guide <https://github.com/kubernetes-sigs/lws/tree/main/docs/examples/vllm>`_ for more details on
|
||||
deploying vLLM on Kubernetes using LWS.
|
||||
Reference in New Issue
Block a user