Commit Graph

107 Commits

Author SHA1 Message Date
Yineng Zhang
429bb0efa2 chore: bump sgl-kernel v0.2.6.post1 (#8200) 2025-07-20 19:50:28 -07:00
Baizhou Zhang
282eb59ff3 Add bf16 output option for dsv3_router_gemm kernel (#7999) 2025-07-20 09:49:37 +08:00
Yineng Zhang
f98e88b9fb chore: bump sgl-kernel v0.2.6 (#8165) 2025-07-19 00:56:18 -07:00
Peng Zhang
c28ad1990d [1/n] chore: decouple quantization implementation from vLLM dependency (#7992) 2025-07-16 15:56:26 -07:00
ykcombat
1ebec1a8b0 [Feature] CUDA Green Context Support (#7649) 2025-07-15 02:49:16 +08:00
Yineng Zhang
8f3173d0b0 chore: bump sgl-kernel v0.2.5 (#7964) 2025-07-11 18:24:20 -07:00
Yineng Zhang
f200af0d8c chore: bump sgl-kernel v0.2.4 (#7800) 2025-07-05 15:03:31 -07:00
Lianmin Zheng
5589b75024 Add treemask mode to build_eagle_tree & release sgl-kernel 0.2.3 (#7756)
Co-authored-by: Pranjal Shankhdhar <pranjal.ssh@gmail.com>
2025-07-05 12:17:05 -07:00
Yineng Zhang
4fece12be9 chore: bump sgl-kernel v0.2.3 (#7784) 2025-07-05 00:05:45 -07:00
SijiaYang
da3890e82a [1/n]: add cutlass W4A8 moe kernel for hopper architecture (#7772)
Signed-off-by: yangsijia.614 <yangsijia.614@bytedance.com>
Co-authored-by: yicwang <yichen.wang@bytedance.com>
2025-07-04 20:50:12 -07:00
Yineng Zhang
aca1101a13 chore: bump sgl-kernel 0.2.2 (#7755) 2025-07-03 12:49:10 -07:00
Yi Zhang
2998c4bdf4 [optimize] fuse renormalize into moe_topk_softmax (#7744)
Co-authored-by: ispobock <ispobaoke@gmail.com>
2025-07-03 12:42:44 -07:00
AniZpZ
8e03b641ba [1/n] apply wna16marlin kernel in moe weight only quantization (#7683)
Co-authored-by: 晟海 <huangtingwei.htw@antgroup.com>
Co-authored-by: yych0745 <1398089567@qq.com>
Co-authored-by: HandH1998 <1335248067@qq.com>
Co-authored-by: 弋云 <yiyun.wyt@antgroup.com>
Co-authored-by: walker-ai <2398833647@qq.com>
2025-07-01 23:21:25 -07:00
Yineng Zhang
637bfee448 chore: bump sgl-kernel v0.2.1 (#7675) 2025-06-30 22:12:33 -07:00
Baizhou Zhang
7248272ccc Add dsv3 router gemm kernel (#7627) 2025-06-29 23:31:55 -07:00
Ke Bao
04b35190e2 Add dsv3 fused a gemm to sgl-kernel (#7630) 2025-06-29 02:52:24 -07:00
Ke Bao
57ab776910 Fuse sorted_token_ids padding to moe_align_block_size kernel (#7437) 2025-06-24 17:44:27 -07:00
Yineng Zhang
e846d95ef6 chore: bump sgl-kernel v0.2.0 (#7490) 2025-06-23 22:29:50 -07:00
Zhiqiang Xie
34c3f9b2d3 kvcache io kernels and test case (#7382) 2025-06-23 11:58:59 -07:00
Lianmin Zheng
55e03b10c4 Fix a bug in BatchTokenIDOut & Misc style and dependency updates (#7457) 2025-06-23 06:20:39 -07:00
Lianmin Zheng
8321f8e45e Release sgl-kernel 0.1.9 (#7232) 2025-06-16 03:37:40 -07:00
Lianmin Zheng
cfceb83d05 Fix sampling for speculative decoding & simplify kernels (#7207) 2025-06-16 03:28:30 -07:00
Yineng Zhang
4473320380 chore: bump v0.1.8.post2 (#7189) 2025-06-14 17:01:48 -07:00
JieXin Liang
ab1a4fa5cb [fix] fix cutlass_mla_backend with cuda_graph and add sm_scale for sgl-kernel cutlass_mla (#7184) 2025-06-14 12:45:41 -07:00
Yineng Zhang
8ab7d93c2e chore: bump v0.1.8.post1 (#7152) 2025-06-13 03:14:26 -07:00
fzyzcjy
5c66c4424f Support new DeepGEMM format in per token group quant (#7146) 2025-06-13 02:00:22 -07:00
fzyzcjy
aa46ed34d2 Remove 200us slow concat kernel (part 1: kernel) (#7145) 2025-06-13 01:58:29 -07:00
Yineng Zhang
7046e0fab7 feat: update blackwell setup (#7119) 2025-06-12 01:54:40 -07:00
Yuan Luo
84727a5139 [sgl-kernel] Add cuda kernel for moe_ep_silu_and_mul (#6919)
Co-authored-by: luoyuan.luo <luoyuan.luo@antgroup.com>
2025-06-11 20:43:08 -07:00
JieXin Liang
18efb5e8e0 [perf][sgl-kernel] extend cutlass_mla_decode to support num_head < 128 (#6929) 2025-06-08 19:37:34 -07:00
Yineng Zhang
6c0a48282a chore: bump sgl-kernel v0.1.7 (#6963) 2025-06-08 02:43:15 -07:00
Yineng Zhang
8db3ac55a9 chore: bump sgl-kernel v0.1.6.post1 (#6955) 2025-06-07 15:25:46 -07:00
Elfie Guo
3e56f557fd Add a CUDA kernel for fusing mapping and weighted sum for MoE. (#6916)
Co-authored-by: Elfie Guo <elfiegxf@gmail.com>
2025-06-07 15:24:39 -07:00
Yineng Zhang
d664ca18f2 chore: bump sgl-kernel v0.1.6 (#6943) 2025-06-07 00:25:22 -07:00
Pavani Majety
0df6765c83 [CUTLASS-FP4-MOE] Introduce CutlassMoEParams class for easy initialization of Cutlass Grouped Gems Metadata (#6887)
Signed-off-by: Pavani Majety <pmajety@nvidia.com>
2025-06-05 13:13:14 -07:00
Yuan Luo
43baba649e [EP] Add cuda kernel for moe_ep_post_reorder (#6837)
Co-authored-by: luoyuan.luo <luoyuan.luo@antgroup.com>
2025-06-05 00:33:47 -07:00
zyksir
8e3797be1c support 1 shot allreduce in 1-node and 2-node using mscclpp (#6277) 2025-06-04 22:11:24 -07:00
Cheng Wan
81964328b7 Set num_fused_shared_experts as num_shared_experts when shared_experts fusion is not disabled (#6736) 2025-06-04 15:53:22 -07:00
Cheng Wan
8a5480528d [Refactor] Rename n_share_experts_fusion as num_fused_shared_experts (#6735) 2025-06-03 17:48:24 -07:00
Pavani Majety
eb38c7d1ca [1/2] Add Kernel support for Cutlass based Fused FP4 MoE (#6093)
Signed-off-by: Pavani Majety <pmajety@nvidia.com>
2025-06-02 13:48:03 -07:00
Yuan Luo
55444ed667 [EP] Add cuda kernel for moe_ep_pre_reorder (#6699)
Co-authored-by: luoyuan.luo <luoyuan.luo@antgroup.com>
2025-06-01 20:49:01 -07:00
Wenxuan Tan
c429919def misc: cache is_hopper_arch (#6799) 2025-06-01 15:28:31 -07:00
Huapeng Zhou
2f7420bc84 [Feat] Enable PDL automatically on Hopper architecture (#5981) 2025-06-01 12:30:17 -07:00
Yineng Zhang
b520d02888 chore: bump sgl-kernel v0.1.5 (#6794) 2025-05-31 14:54:00 -07:00
Yineng Zhang
d71f3f0a2a chore: bump sgl-kernel v0.1.4 (#6522) 2025-05-22 09:47:42 -07:00
HandH1998
4d643f6c7a [1/2] Support Qserve (#6457)
Co-authored-by: yych0745 <1398089567@qq.com>
Co-authored-by: sleepcoo <sleepcoo@gmail.com>
2025-05-21 19:48:59 -07:00
Yineng Zhang
3d7f7a43c8 chore: bump sgl-kernel v0.1.3 (#6368) 2025-05-17 00:15:55 -07:00
Elfie Guo
6fc9357503 [2/2] Add python wrapper for CUTLASS FP8 Blockscale MoE Kernel. (#5694) 2025-05-16 13:14:07 -07:00
Lianmin Zheng
e8e18dcdcc Revert "fix some typos" (#6244) 2025-05-12 12:53:26 -07:00
applesaucethebun
d738ab52f8 fix some typos (#6209)
Co-authored-by: Brayden Zhong <b8zhong@uwaterloo.ca>
2025-05-13 01:42:38 +08:00