This website requires JavaScript.
Explore
Help
Register
Sign In
EngineX
/
xc-llm-kunlun
Watch
3
Star
0
Fork
0
You've already forked xc-llm-kunlun
Code
Issues
Pull Requests
Actions
Projects
Releases
Wiki
Activity
Files
eb40e8a07a915ed37758edb4ac5bc3071445106f
xc-llm-kunlun
/
vllm_kunlun
/
ops
/
quantization
History
baoqian426
eb40e8a07a
[Bugfix] fix can not import compressed_tensors (
#87
)
...
Co-authored-by: root <
root@rdtest-node1150.bcc-zwlt.baidu.com
>
2026-01-07 11:32:10 +08:00
..
compressed_tensors
[Bugfix] fix can not import compressed_tensors (
#87
)
2026-01-07 11:32:10 +08:00
kernels
[fix]update compressed-tensors scheme
2026-01-06 22:30:27 +08:00
__init__.py
Initial commit for vLLM-Kunlun Plugin
2025-12-10 12:05:39 +08:00
awq.py
[dev] support AWQ/GPTQ quantization for dense models
2025-12-24 13:46:06 +08:00
gptq.py
[dev] support AWQ/GPTQ quantization for dense models
2025-12-24 13:46:06 +08:00