Li Wei
|
71bd70ad6c
|
[Feature] support compressed-tensors w4a16 quantization (#154)
- native int4 kimi model inference is supported
Signed-off-by: Li Wei <liwei.109@outlook.com>
|
2026-01-27 19:56:22 +08:00 |
|
Shiwen Tang
|
0711c1abfa
|
[Feature] Support AWQ MoE W4A16 Quantization (#142)
Signed-off-by: tangshiwen <tangshiwen@baidu.com>
Co-authored-by: Li Wei <liwei.109@outlook.com>
|
2026-01-26 18:56:05 +08:00 |
|
Li Wei
|
c403d921ff
|
[doc] update quantization guide doc (#88)
|
2026-01-07 15:39:51 +08:00 |
|
chenyili
|
7c22d621fb
|
提交vllm0.11.0开发分支
|
2025-12-10 17:51:24 +08:00 |
|
dongxinyu03
|
c728e52505
|
Initial commit for vLLM-Kunlun Plugin
|
2025-12-10 12:05:39 +08:00 |
|