Logo
Explore Help
Register Sign In
EngineX/xc-llm-ascend
3
0
Fork 0
You've already forked xc-llm-ascend
Code Issues Pull Requests Projects Releases Wiki Activity
Files
c85cc045f893293e3b44e24d2e1f01ddc5849ea8
xc-llm-ascend/vllm_ascend/ops/fused_moe
History
zzzzwwjj 71f729a661 Revert "moe_gating_top_k" (#5512)
Reverts vllm-project/vllm-ascend#5271

It breaks e2e test

- vLLM version: v0.13.0
- vLLM main:
45c1ca1ca1
2025-12-30 15:05:47 +08:00
..
__init__.py
[Refactor] [MoE] Rename moe-related classes & files (#3646)
2025-10-25 11:22:03 +08:00
comm_utils.py
[Refactor] [MoE] Rename moe-related classes & files (#3646)
2025-10-25 11:22:03 +08:00
experts_selector.py
Revert "moe_gating_top_k" (#5512)
2025-12-30 15:05:47 +08:00
fused_moe.py
[EPLB][refactor] Modification of the initialization logic for expert_map and log2phy(depend on pr5285) (#5311)
2025-12-29 09:26:14 +08:00
moe_comm_method.py
[Feature]Use DispatchGmmCombineDecode operator to replace MC2(Optional) (#5040)
2025-12-21 15:23:59 +08:00
moe_mlp.py
[refactor] Remove unnecessary attributes from set_ascend_forward_context (#5204)
2025-12-23 08:49:52 +08:00
prepare_finalize.py
[Refactor][MoE] Reuse vLLM's all_reduce logic (#5189)
2025-12-23 18:53:48 +08:00
token_dispatcher.py
Revert "[feat] enable hierarchical mc2 ops on A2 by default (#5300)" (#5434)
2025-12-27 17:06:58 +08:00
Powered by Gitea Version: 1.24.3 Page: 186ms Template: 6ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API