[RL] allow weight updation with dp attention enabled (#6311)

This commit is contained in:
Zilin Zhu
2025-05-21 16:58:55 +08:00
committed by GitHub
parent 669caa0a3f
commit 7c347259ff

View File

@@ -893,8 +893,8 @@ class TokenizerManager:
) -> Tuple[bool, str]:
self.auto_create_handle_loop()
assert (
self.server_args.dp_size == 1
), "dp_size must be 1 for update weights from distributed"
self.server_args.dp_size == 1 or self.server_args.enable_dp_attention
), "dp_size must be 1 or dp attention must be enabled for update weights from distributed"
# This means that weight sync
# cannot run while requests are in progress.
@@ -909,8 +909,8 @@ class TokenizerManager:
) -> Tuple[bool, str]:
self.auto_create_handle_loop()
assert (
self.server_args.dp_size == 1
), "dp_size must be 1 for update weights from distributed"
self.server_args.dp_size == 1 or self.server_args.enable_dp_attention
), "dp_size must be 1 or dp attention must be enabled for update weights from tensor"
# This means that weight sync
# cannot run while requests are in progress.