Logo
Explore Help
Register Sign In
EngineX/xc-llm-ascend
3
0
Fork 0
You've already forked xc-llm-ascend
Code Issues Pull Requests Actions Projects Releases Wiki Activity
Files
b6d63bbd52945192bba2ad8038e6c3bc241bc3b5
xc-llm-ascend/vllm_ascend/ops/moe
History
wangxiyuan a2e4c3fe78 Revert "[cherry-pick][refactor]support gatingtopk operator generalization (#4050)" (#4352)
This reverts commit c87a77e8b4.

it breaks ops e2e test

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
2025-11-21 23:03:20 +08:00
..
__init__.py
[main] [refactor] refactor common_fused_moe.py (#2706)
2025-09-08 20:09:50 +08:00
comm_utils.py
Refactor tensor_parallel and comm_utils (#2814)
2025-09-11 21:26:36 +08:00
experts_selector.py
Revert "[cherry-pick][refactor]support gatingtopk operator generalization (#4050)" (#4352)
2025-11-21 23:03:20 +08:00
fused_moe_prepare_and_finalize.py
[v0.11.0] [Bugfix] [MoE]fix error in deepseek when using allgather (#3827)
2025-10-30 14:59:46 +08:00
moe_comm_method.py
[BugFix]Fix group list type of mc2. (#3890)
2025-10-30 21:44:14 +08:00
moe_mlp.py
[Feat]Qwen3 Moe supports npu_add_rms_norm_quant op by default, update op with bias, resolve conflict with weight prefetch (#3465)
2025-10-17 09:30:51 +08:00
token_dispatcher.py
[0.11.0-dev][Bugfix][EPLB] Quick fix for missing log2phy conversion (#4150)
2025-11-13 14:32:40 +08:00
Powered by Gitea Version: 1.24.3 Page: 255ms Template: 7ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API