Update vllm version to support llama3.1 (#705)
This commit is contained in:
@@ -222,6 +222,7 @@ def launch_server(
|
||||
detokenizer_port=ports[2],
|
||||
nccl_ports=ports[3:],
|
||||
)
|
||||
logger.info(f"{server_args=}")
|
||||
|
||||
# Handle multi-node tensor parallelism
|
||||
if server_args.nnodes > 1:
|
||||
|
||||
Reference in New Issue
Block a user