From c93293c57e16566c18443bae1bfacc107b714c74 Mon Sep 17 00:00:00 2001 From: Ikko Eltociear Ashimine Date: Wed, 10 Apr 2024 00:39:30 +0900 Subject: [PATCH] Update README.md (#358) --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index b388df8d9..018d1f008 100644 --- a/README.md +++ b/README.md @@ -362,7 +362,7 @@ python -m sglang.launch_server --model-path meta-llama/Llama-2-7b-chat-hf --port ``` python -m sglang.launch_server --model-path meta-llama/Llama-2-7b-chat-hf --port 30000 --mem-fraction-static 0.7 ``` -- You can turn on [flashinfer](docs/flashinfer.md) to acclerate the inference by using highly optimized CUDA kernels. +- You can turn on [flashinfer](docs/flashinfer.md) to accelerate the inference by using highly optimized CUDA kernels. ### Supported Models - Llama