Commit Graph

10 Commits

Author SHA1 Message Date
baoqian426
ee0f50e68f [Feature] support deepseek v3/r1/v3.2 (#78)
* [Feature] support deepseek v3/r1/v3.2

* fix gpt_oss

* update readme

* update readme

---------

Co-authored-by: hanhaowen <hanhaowen@baidu.com>
2026-01-05 22:55:35 +08:00
hanhaowen
b015bb76fd remove qwen2.py llama.py fix llama output 2025-12-31 11:39:37 +08:00
Xinyu Dong
b3c30a3cb9 [Feature] Support XiaoMi MIMO Flash V2 (#62)
* [Feature] Support MIMO Flash V2
2025-12-31 10:16:33 +08:00
ldh2020
58c1db5073 [Bugfix] fix the bug of the flash_attention in Qwen3-Next 2025-12-21 10:34:43 +08:00
chenyili0619
2e2933d217 [Bug] Fixed the issue where an error occurred when the request included a seed. 2025-12-18 13:03:34 +08:00
baoqian426
fae22c2e62 Merge pull request #3 from xyDong0223/main
[Kernel] Enable fast random sample on Kunlun3 Platform
2025-12-11 11:47:30 +08:00
xyDong0223
af2cd6097f [Kernell] fix miss import os 2025-12-11 11:17:28 +08:00
xyDong0223
670c2397b8 [Kernel] Enable fast random sample on Kunlun P 2025-12-10 21:52:48 +08:00
chenyili
7c22d621fb 提交vllm0.11.0开发分支 2025-12-10 17:51:24 +08:00
dongxinyu03
c728e52505 Initial commit for vLLM-Kunlun Plugin 2025-12-10 12:05:39 +08:00