This website requires JavaScript.
Explore
Help
Register
Sign In
EngineX
/
xc-llm-kunlun
Watch
3
Star
0
Fork
0
You've already forked xc-llm-kunlun
Code
Issues
Pull Requests
Projects
Releases
Wiki
Activity
Files
f2019b145f0ca50153640486b4d9ed8cb9b7a87b
xc-llm-kunlun
/
vllm_kunlun
/
v1
/
attention
/
backends
History
baoqian426
2512259944
longcontext chunk make attention crash, fix it (
#117
)
...
Co-authored-by: root <
root@rdtest-node1150.bcc-zwlt.baidu.com
>
2026-01-17 18:38:23 +08:00
..
mla
longcontext chunk make attention crash, fix it (
#117
)
2026-01-17 18:38:23 +08:00
__init__.py
Initial commit for vLLM-Kunlun Plugin
2025-12-10 12:05:39 +08:00
kunlun_attn.py
remove qwen2.py llama.py fix llama output
2025-12-31 11:39:37 +08:00