[RL] allow weight updation with dp attention enabled (#6311)
This commit is contained in:
@@ -893,8 +893,8 @@ class TokenizerManager:
|
|||||||
) -> Tuple[bool, str]:
|
) -> Tuple[bool, str]:
|
||||||
self.auto_create_handle_loop()
|
self.auto_create_handle_loop()
|
||||||
assert (
|
assert (
|
||||||
self.server_args.dp_size == 1
|
self.server_args.dp_size == 1 or self.server_args.enable_dp_attention
|
||||||
), "dp_size must be 1 for update weights from distributed"
|
), "dp_size must be 1 or dp attention must be enabled for update weights from distributed"
|
||||||
|
|
||||||
# This means that weight sync
|
# This means that weight sync
|
||||||
# cannot run while requests are in progress.
|
# cannot run while requests are in progress.
|
||||||
@@ -909,8 +909,8 @@ class TokenizerManager:
|
|||||||
) -> Tuple[bool, str]:
|
) -> Tuple[bool, str]:
|
||||||
self.auto_create_handle_loop()
|
self.auto_create_handle_loop()
|
||||||
assert (
|
assert (
|
||||||
self.server_args.dp_size == 1
|
self.server_args.dp_size == 1 or self.server_args.enable_dp_attention
|
||||||
), "dp_size must be 1 for update weights from distributed"
|
), "dp_size must be 1 or dp attention must be enabled for update weights from tensor"
|
||||||
|
|
||||||
# This means that weight sync
|
# This means that weight sync
|
||||||
# cannot run while requests are in progress.
|
# cannot run while requests are in progress.
|
||||||
|
|||||||
Reference in New Issue
Block a user