Logo
Explore Help
Register Sign In
EngineX/xc-llm-ascend
3
0
Fork 0
You've already forked xc-llm-ascend
Code Issues Pull Requests Projects Releases Wiki Activity
Files
2ee17e50a1525abc0d47559571cbae5d41b662b3
xc-llm-ascend/vllm_ascend/ops/fused_moe
History
zzzzwwjj 71f729a661 Revert "moe_gating_top_k" (#5512)
Reverts vllm-project/vllm-ascend#5271

It breaks e2e test

- vLLM version: v0.13.0
- vLLM main:
45c1ca1ca1
2025-12-30 15:05:47 +08:00
..
__init__.py
[Refactor] [MoE] Rename moe-related classes & files (#3646)
2025-10-25 11:22:03 +08:00
comm_utils.py
[Refactor] [MoE] Rename moe-related classes & files (#3646)
2025-10-25 11:22:03 +08:00
experts_selector.py
Revert "moe_gating_top_k" (#5512)
2025-12-30 15:05:47 +08:00
fused_moe.py
[EPLB][refactor] Modification of the initialization logic for expert_map and log2phy(depend on pr5285) (#5311)
2025-12-29 09:26:14 +08:00
moe_comm_method.py
[Feature]Use DispatchGmmCombineDecode operator to replace MC2(Optional) (#5040)
2025-12-21 15:23:59 +08:00
moe_mlp.py
[refactor] Remove unnecessary attributes from set_ascend_forward_context (#5204)
2025-12-23 08:49:52 +08:00
prepare_finalize.py
[Refactor][MoE] Reuse vLLM's all_reduce logic (#5189)
2025-12-23 18:53:48 +08:00
token_dispatcher.py
Revert "[feat] enable hierarchical mc2 ops on A2 by default (#5300)" (#5434)
2025-12-27 17:06:58 +08:00
Powered by Gitea Version: 1.24.3 Page: 321ms Template: 66ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API