[6/N] MoE Refactor: Cleanup MoE-related configs (#8849)
This commit is contained in:
@@ -2413,7 +2413,7 @@ def require_mlp_tp_gather(server_args):
|
||||
return True
|
||||
elif not server_args.enable_dp_lm_head:
|
||||
return True
|
||||
elif server_args.moe_a2a_backend is None:
|
||||
elif server_args.moe_a2a_backend == "none":
|
||||
return True
|
||||
else:
|
||||
return (
|
||||
@@ -2429,7 +2429,7 @@ def require_attn_tp_gather(server_args):
|
||||
Check if the input of attention is scattered.
|
||||
"""
|
||||
assert server_args.moe_dense_tp_size in [1, None]
|
||||
if server_args.moe_a2a_backend is not None or server_args.moe_dense_tp_size == 1:
|
||||
if server_args.moe_a2a_backend != "none" or server_args.moe_dense_tp_size == 1:
|
||||
if server_args.enable_dp_attention:
|
||||
return server_args.dp_size < server_args.tp_size
|
||||
else:
|
||||
|
||||
Reference in New Issue
Block a user