[Bug] Fix no apply_top_k_top_p issue. (#101)

This commit is contained in:
Jin Hanyu
2026-01-12 16:38:03 +08:00
committed by GitHub
parent 18fc1c006e
commit bd90350968

View File

@@ -11,6 +11,7 @@ from vllm import envs
from vllm.logger import init_logger
from vllm.platforms import current_platform
import xtorch_ops
import os
logger = init_logger(__name__)
@@ -28,6 +29,7 @@ class TopKTopPSampler(nn.Module):
logger.info_once(
"Using FlashInfer for top-p & top-k sampling.")
self.forward = self.forward_kunlun
self.apply_top_k_top_p = apply_top_k_top_p
def forward_native(
self,