This website requires JavaScript.
Explore
Help
Register
Sign In
EngineX
/
xc-llm-kunlun
Watch
3
Star
0
Fork
0
You've already forked xc-llm-kunlun
Code
Issues
Pull Requests
Actions
Projects
Releases
Wiki
Activity
Files
main
xc-llm-kunlun
/
vllm_kunlun
/
ops
/
quantization
/
kernels
History
Li Wei
71bd70ad6c
[Feature] support compressed-tensors w4a16 quantization (
#154
)
...
- native int4 kimi model inference is supported Signed-off-by: Li Wei <
liwei.109@outlook.com
>
2026-01-27 19:56:22 +08:00
..
__init__.py
[Feature] support deepseek v3/r1/v3.2 (
#78
)
2026-01-05 22:55:35 +08:00
kunlun_exllama_linear.py
[Feature] support compressed-tensors w4a16 quantization (
#154
)
2026-01-27 19:56:22 +08:00
kunlun_scale_mm.py
[Feature] support compressed-tensors w4a16 quantization (
#154
)
2026-01-27 19:56:22 +08:00
quant_ops.py
[Feature] support compressed-tensors w4a16 quantization (
#154
)
2026-01-27 19:56:22 +08:00