Logo
Explore Help
Register Sign In
EngineX/xc-llm-ascend
3
0
Fork 0
You've already forked xc-llm-ascend
Code Issues Pull Requests Projects Releases Wiki Activity
Files
ceadc2788da2a3d726d644e27aeaef14c6966405
xc-llm-ascend/vllm_ascend/ops/moe
History
1092626063 ceadc2788d Revert "[refactor]support gatingtopk operator generalization (#4356)" (#4873)
This reverts commit c4a11a745a.

ops npu_gating_top_k caused Qwen3-30B precision problem, so revert it.

Signed-off-by: 1092626063 <1092626063@qq.com>
2025-12-10 15:45:20 +08:00
..
__init__.py
[main] [refactor] refactor common_fused_moe.py (#2706)
2025-09-08 20:09:50 +08:00
comm_utils.py
Refactor tensor_parallel and comm_utils (#2814)
2025-09-11 21:26:36 +08:00
experts_selector.py
Revert "[refactor]support gatingtopk operator generalization (#4356)" (#4873)
2025-12-10 15:45:20 +08:00
fused_moe_prepare_and_finalize.py
[v0.11.0] [Bugfix] [MoE]fix error in deepseek when using allgather (#3827)
2025-10-30 14:59:46 +08:00
moe_comm_method.py
[BugFix]Fix group list type of mc2. (#3890)
2025-10-30 21:44:14 +08:00
moe_mlp.py
BugFix: Resolve shape mismatch in eplb update and calculation issues in quant_apply_mlp (#4777)
2025-12-09 15:46:58 +08:00
token_dispatcher.py
BugFix: Resolve shape mismatch in eplb update and calculation issues in quant_apply_mlp (#4777)
2025-12-09 15:46:58 +08:00
Powered by Gitea Version: 1.24.3 Page: 274ms Template: 7ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API