[Bug] Fixed the issue where an error occurred when the request included a seed.
This commit is contained in:
@@ -42,7 +42,7 @@ class TopKTopPSampler(nn.Module):
|
||||
"""
|
||||
logits = apply_top_k_top_p(logits, k, p)
|
||||
probs = logits.softmax(dim=-1, dtype=torch.float32)
|
||||
return random_sample(probs, generators)
|
||||
return random_sample(probs, generators), None
|
||||
|
||||
def forward_kunlun(
|
||||
self,
|
||||
@@ -199,4 +199,4 @@ def flashinfer_sample(
|
||||
next_token_ids = xtorch_ops.top_k_top_p_sampling_from_probs(
|
||||
probs, top_k=k, top_p=p, deterministic=True)
|
||||
|
||||
return next_token_ids.view(-1)
|
||||
return next_token_ids.view(-1)
|
||||
|
||||
Reference in New Issue
Block a user