Merge pull request #29 from chenyili0619/28-v0110-only-enble-top-p-or-k-occur-error
[Bug] Fixed the issue where an error occurred when the request includ…
This commit is contained in:
@@ -42,7 +42,7 @@ class TopKTopPSampler(nn.Module):
|
|||||||
"""
|
"""
|
||||||
logits = apply_top_k_top_p(logits, k, p)
|
logits = apply_top_k_top_p(logits, k, p)
|
||||||
probs = logits.softmax(dim=-1, dtype=torch.float32)
|
probs = logits.softmax(dim=-1, dtype=torch.float32)
|
||||||
return random_sample(probs, generators)
|
return random_sample(probs, generators), None
|
||||||
|
|
||||||
def forward_kunlun(
|
def forward_kunlun(
|
||||||
self,
|
self,
|
||||||
@@ -199,4 +199,4 @@ def flashinfer_sample(
|
|||||||
next_token_ids = xtorch_ops.top_k_top_p_sampling_from_probs(
|
next_token_ids = xtorch_ops.top_k_top_p_sampling_from_probs(
|
||||||
probs, top_k=k, top_p=p, deterministic=True)
|
probs, top_k=k, top_p=p, deterministic=True)
|
||||||
|
|
||||||
return next_token_ids.view(-1)
|
return next_token_ids.view(-1)
|
||||||
|
|||||||
Reference in New Issue
Block a user