This website requires JavaScript.
Explore
Help
Register
Sign In
EngineX
/
xc-llm-kunlun
Watch
3
Star
0
Fork
0
You've already forked xc-llm-kunlun
Code
Issues
Pull Requests
Actions
Projects
Releases
Wiki
Activity
Files
main
xc-llm-kunlun
/
docs
/
source
/
user_guide
/
feature_guide
History
Li Wei
71bd70ad6c
[Feature] support compressed-tensors w4a16 quantization (
#154
)
...
- native int4 kimi model inference is supported Signed-off-by: Li Wei <
liwei.109@outlook.com
>
2026-01-27 19:56:22 +08:00
..
graph_mode.md
提交vllm0.11.0开发分支
2025-12-10 17:51:24 +08:00
index.md
Initial commit for vLLM-Kunlun Plugin
2025-12-10 12:05:39 +08:00
lora.md
Initial commit for vLLM-Kunlun Plugin
2025-12-10 12:05:39 +08:00
quantization.md
[Feature] support compressed-tensors w4a16 quantization (
#154
)
2026-01-27 19:56:22 +08:00