Sync from v0.13
This commit is contained in:
9
docs/usage/README.md
Normal file
9
docs/usage/README.md
Normal file
@@ -0,0 +1,9 @@
|
||||
# Using vLLM
|
||||
|
||||
First, vLLM must be [installed](../getting_started/installation/README.md) for your chosen device in either a Python or Docker environment.
|
||||
|
||||
Then, vLLM supports the following usage patterns:
|
||||
|
||||
- [Inference and Serving](../serving/offline_inference.md): Run a single instance of a model.
|
||||
- [Deployment](../deployment/docker.md): Scale up model instances for production.
|
||||
- [Training](../training/rlhf.md): Train or fine-tune a model.
|
||||
Reference in New Issue
Block a user