fix bmm fp8 (#4926)
This commit is contained in:
@@ -82,7 +82,10 @@ TORCH_LIBRARY_EXPAND(sgl_kernel, m) {
|
|||||||
/*
|
/*
|
||||||
* From FlashInfer
|
* From FlashInfer
|
||||||
*/
|
*/
|
||||||
m.def("bmm_fp8", bmm_fp8);
|
m.def(
|
||||||
|
"bmm_fp8(Tensor A, Tensor B, Tensor! D, Tensor A_scale, Tensor B_scale, Tensor workspace_buffer, int "
|
||||||
|
"cublas_handle, int cuda_stream) -> ()");
|
||||||
|
m.impl("bmm_fp8", torch::kCUDA, &bmm_fp8);
|
||||||
m.def("min_p_sampling_from_probs", min_p_sampling_from_probs);
|
m.def("min_p_sampling_from_probs", min_p_sampling_from_probs);
|
||||||
m.def("top_k_renorm_probs", top_k_renorm_probs);
|
m.def("top_k_renorm_probs", top_k_renorm_probs);
|
||||||
m.def("top_p_renorm_probs", top_p_renorm_probs);
|
m.def("top_p_renorm_probs", top_p_renorm_probs);
|
||||||
|
|||||||
Reference in New Issue
Block a user