This website requires JavaScript.
Explore
Help
Register
Sign In
EngineX
/
xc-llm-ascend
Watch
3
Star
0
Fork
0
You've already forked xc-llm-ascend
Code
Issues
Pull Requests
Actions
Projects
Releases
Wiki
Activity
Files
1715230867048aaf3102dbe6448b3c476db74c9e
xc-llm-ascend
/
vllm_ascend
History
HongtaoYang
1715230867
[CI] Upgrade to newest pta.(MLA and FusedMoE) (
#189
)
...
Upgrade to newest pta.(MLA and FusedMoE) --------- Signed-off-by: SidaoY <
1024863041@qq.com
>
2025-02-27 18:50:52 +08:00
..
ops
[CI] Upgrade to newest pta.(MLA and FusedMoE) (
#189
)
2025-02-27 18:50:52 +08:00
quantization
[Core] Cherry pick from 0.7.1 to keep the main code newest (
#127
)
2025-02-21 17:07:37 +08:00
__init__.py
[Core] Init vllm-ascend (
#3
)
2025-02-05 10:53:12 +08:00
attention.py
[CI] Upgrade to newest pta.(MLA and FusedMoE) (
#189
)
2025-02-27 18:50:52 +08:00
communicator.py
[dist] revert communicator patch (
#66
)
2025-02-17 11:42:33 +08:00
model_runner.py
[ModelRunner] Fix cuda hard code in model runner (
#155
)
2025-02-27 14:16:46 +08:00
platform.py
Ray Worker Ops Optimization (
#136
)
2025-02-21 22:45:15 +08:00
utils.py
[Worker] Register mindie_turbo while initializing NPUWorker (
#13
)
2025-02-07 16:47:17 +08:00
worker.py
[Worker]Lazy import torch_npu (
#184
)
2025-02-27 16:52:11 +08:00