Update vllm version to support llama3.1 (#705)

This commit is contained in:
Ying Sheng
2024-07-23 13:49:34 -07:00
committed by GitHub
parent fa7ccb3316
commit 444a02441a
4 changed files with 5 additions and 9 deletions

View File

@@ -222,6 +222,7 @@ def launch_server(
detokenizer_port=ports[2],
nccl_ports=ports[3:],
)
logger.info(f"{server_args=}")
# Handle multi-node tensor parallelism
if server_args.nnodes > 1: