[6/N] MoE Refactor: Cleanup MoE-related configs (#8849)
This commit is contained in:
@@ -25,7 +25,6 @@ import torch
|
||||
import torch.distributed
|
||||
|
||||
from sglang.srt.eplb.expert_location import ExpertLocationMetadata
|
||||
from sglang.srt.managers.schedule_batch import global_server_args_dict
|
||||
from sglang.srt.model_executor.forward_batch_info import ForwardBatch
|
||||
from sglang.srt.server_args import ServerArgs
|
||||
from sglang.srt.utils import Withable, get_bool_env_var
|
||||
@@ -288,14 +287,14 @@ class _SinglePassGatherer(ABC):
|
||||
)
|
||||
|
||||
if server_args.expert_distribution_recorder_mode == "stat_approx":
|
||||
if server_args.moe_a2a_backend is not None and (
|
||||
if server_args.moe_a2a_backend != "none" and (
|
||||
server_args.deepep_mode == "normal"
|
||||
):
|
||||
return _DeepepNormalSinglePassGatherer(expert_location_metadata, rank)
|
||||
else:
|
||||
raise NotImplementedError
|
||||
|
||||
if server_args.moe_a2a_backend is not None:
|
||||
if server_args.moe_a2a_backend != "none":
|
||||
if server_args.deepep_mode == "normal":
|
||||
return _SelectExpertsSinglePassGatherer(expert_location_metadata, rank)
|
||||
elif server_args.deepep_mode == "low_latency":
|
||||
|
||||
Reference in New Issue
Block a user