Logo
Explore Help
Register Sign In
EngineX/xc-llm-ascend
3
0
Fork 0
You've already forked xc-llm-ascend
Code Issues Pull Requests Actions Projects Releases Wiki Activity
Files
593a96056cfaaa64746ff05f3da526382a3d5d64
xc-llm-ascend/vllm_ascend/ops/moe
History
wangxiyuan a2e4c3fe78 Revert "[cherry-pick][refactor]support gatingtopk operator generalization (#4050)" (#4352)
This reverts commit c87a77e8b4.

it breaks ops e2e test

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
2025-11-21 23:03:20 +08:00
..
__init__.py
[main] [refactor] refactor common_fused_moe.py (#2706)
2025-09-08 20:09:50 +08:00
comm_utils.py
Refactor tensor_parallel and comm_utils (#2814)
2025-09-11 21:26:36 +08:00
experts_selector.py
Revert "[cherry-pick][refactor]support gatingtopk operator generalization (#4050)" (#4352)
2025-11-21 23:03:20 +08:00
fused_moe_prepare_and_finalize.py
[v0.11.0] [Bugfix] [MoE]fix error in deepseek when using allgather (#3827)
2025-10-30 14:59:46 +08:00
moe_comm_method.py
[BugFix]Fix group list type of mc2. (#3890)
2025-10-30 21:44:14 +08:00
moe_mlp.py
[Feat]Qwen3 Moe supports npu_add_rms_norm_quant op by default, update op with bias, resolve conflict with weight prefetch (#3465)
2025-10-17 09:30:51 +08:00
token_dispatcher.py
[0.11.0-dev][Bugfix][EPLB] Quick fix for missing log2phy conversion (#4150)
2025-11-13 14:32:40 +08:00
Powered by Gitea Version: 1.24.3 Page: 75ms Template: 6ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API