Logo
Explore Help
Register Sign In
EngineX/xc-llm-ascend
3
0
Fork 0
You've already forked xc-llm-ascend
Code Issues Pull Requests Actions Projects Releases Wiki Activity
Files
6eddbd2521d9f22b90b302d65f986254251d49cb
xc-llm-ascend/vllm_ascend/ops
History
wangxiyuan f6e5decc10 [CI] upgrade to vllm 0.9.0 (#959)
Upgrade to vllm 0.9.0.
0.8.5 will not be supported any more.

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
2025-05-28 21:18:41 +08:00
..
__init__.py
[MISC] Clean up torch_npu (#688)
2025-04-29 18:03:38 +08:00
activation.py
Optimize qwen2_vl and qwen2_5_vl (#701)
2025-04-30 14:22:38 +08:00
attention.py
[Fix] Set div_mode to False and fix view_as position (#912)
2025-05-22 09:57:25 +08:00
cache.py
port deepseekv2 and mtp to main branch (#429)
2025-04-19 17:38:18 +08:00
common_fused_moe.py
[Attention][Kernel]moe support for llama4 and mllama4 (#740)
2025-05-13 19:12:40 +08:00
fused_moe.py
[CI] upgrade to vllm 0.9.0 (#959)
2025-05-28 21:18:41 +08:00
layernorm.py
[CI]Add model basic accuracy test(Qwen2.5-0.5B-Instruct) (#460)
2025-04-17 14:59:56 +08:00
rotary_embedding.py
[Bugfix] Correct method call for _set_cos_sin_cache (#774)
2025-05-09 12:55:57 +08:00
vocab_parallel_embedding.py
port deepseekv2 and mtp to main branch (#429)
2025-04-19 17:38:18 +08:00
Powered by Gitea Version: 1.24.3 Page: 80ms Template: 6ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API