Logo
Explore Help
Register Sign In
EngineX/xc-llm-ascend
3
0
Fork 0
You've already forked xc-llm-ascend
Code Issues Pull Requests Projects Releases Wiki Activity
Files
073097a9a1ddf8c81737ed9f9f9fe36264884327
xc-llm-ascend/vllm_ascend/ops/fused_moe
History
zzzzwwjj 71f729a661 Revert "moe_gating_top_k" (#5512)
Reverts vllm-project/vllm-ascend#5271

It breaks e2e test

- vLLM version: v0.13.0
- vLLM main:
45c1ca1ca1
2025-12-30 15:05:47 +08:00
..
__init__.py
[Refactor] [MoE] Rename moe-related classes & files (#3646)
2025-10-25 11:22:03 +08:00
comm_utils.py
[Refactor] [MoE] Rename moe-related classes & files (#3646)
2025-10-25 11:22:03 +08:00
experts_selector.py
Revert "moe_gating_top_k" (#5512)
2025-12-30 15:05:47 +08:00
fused_moe.py
[EPLB][refactor] Modification of the initialization logic for expert_map and log2phy(depend on pr5285) (#5311)
2025-12-29 09:26:14 +08:00
moe_comm_method.py
[Feature]Use DispatchGmmCombineDecode operator to replace MC2(Optional) (#5040)
2025-12-21 15:23:59 +08:00
moe_mlp.py
[refactor] Remove unnecessary attributes from set_ascend_forward_context (#5204)
2025-12-23 08:49:52 +08:00
prepare_finalize.py
[Refactor][MoE] Reuse vLLM's all_reduce logic (#5189)
2025-12-23 18:53:48 +08:00
token_dispatcher.py
Revert "[feat] enable hierarchical mc2 ops on A2 by default (#5300)" (#5434)
2025-12-27 17:06:58 +08:00
Powered by Gitea Version: 1.24.3 Page: 340ms Template: 9ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API