[Bugfix] mooncake connector support external dp & update readme (#3579)
### What this PR does / why we need it? mooncake connector support external dp & update readme ### Does this PR introduce _any_ user-facing change? ### How was this patch tested? - vLLM version: v0.11.0rc3 - vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0 --------- Signed-off-by: liziyu <liziyu16@huawei.com>
This commit is contained in:
@@ -701,7 +701,7 @@ class MooncakeConnectorScheduler:
|
||||
# Handshake base port
|
||||
self.side_channel_port = (
|
||||
vllm_config.kv_transfer_config.kv_port +
|
||||
vllm_config.parallel_config.data_parallel_rank_local *
|
||||
vllm_config.parallel_config.data_parallel_rank *
|
||||
vllm_config.parallel_config.tensor_parallel_size)
|
||||
|
||||
# Requests that need to start recv.
|
||||
@@ -891,7 +891,7 @@ class MooncakeConnectorWorker:
|
||||
self.tp_rank = get_tensor_model_parallel_rank()
|
||||
self.tp_size = vllm_config.parallel_config.tensor_parallel_size
|
||||
self.tp_group = get_tp_group()
|
||||
self.dp_rank = vllm_config.parallel_config.data_parallel_rank_local
|
||||
self.dp_rank = vllm_config.parallel_config.data_parallel_rank
|
||||
self.dp_size = vllm_config.parallel_config.data_parallel_size_local
|
||||
self.kv_caches: dict[str, torch.Tensor] = {}
|
||||
self.side_channel_host = get_ip()
|
||||
@@ -902,7 +902,7 @@ class MooncakeConnectorWorker:
|
||||
# Handshake base port
|
||||
self.side_channel_port = (
|
||||
vllm_config.kv_transfer_config.kv_port +
|
||||
vllm_config.parallel_config.data_parallel_rank_local *
|
||||
vllm_config.parallel_config.data_parallel_rank *
|
||||
vllm_config.parallel_config.tensor_parallel_size)
|
||||
self.handshake_port = self.side_channel_port + self.tp_rank
|
||||
self.sockets: dict = {}
|
||||
|
||||
Reference in New Issue
Block a user