Support penalty in overlap mode; return logprob with chunked prefill; improve benchmark scripts (#3988)
Co-authored-by: SangBin Cho <rkooo567@gmail.com> Co-authored-by: dhou-xai <dhou@x.ai> Co-authored-by: Hanming Lu <hanming_lu@berkeley.edu>
This commit is contained in:
@@ -168,9 +168,9 @@ def _run_subprocess(
|
||||
hf_tokenizer = get_tokenizer(model_path, trust_remote_code=True)
|
||||
|
||||
hf_outputs = HFRunner.forward_generation_raw(
|
||||
base_model=hf_model,
|
||||
prompts=_PROMPTS,
|
||||
max_new_tokens=_MAX_NEW_TOKENS,
|
||||
base_model=hf_model,
|
||||
tokenizer=hf_tokenizer,
|
||||
lora_paths=None,
|
||||
torch_dtype=_TORCH_DTYPE,
|
||||
|
||||
Reference in New Issue
Block a user