Offline LLM Engine Benchmark Throughput (#1968)

Co-authored-by: ByronHsu <byronhsu1230@gmail.com>
This commit is contained in:
zolinthecow
2024-11-14 21:59:33 -08:00
committed by GitHub
parent ea53c63bad
commit f6dd648620
4 changed files with 358 additions and 31 deletions

View File

@@ -768,7 +768,7 @@ class Engine:
self,
# The input prompt. It can be a single prompt or a batch of prompts.
prompt: Optional[Union[List[str], str]] = None,
sampling_params: Optional[Dict] = None,
sampling_params: Optional[Union[List[Dict], Dict]] = None,
# The token ids for text; one can either specify text or input_ids.
input_ids: Optional[Union[List[List[int]], List[int]]] = None,
return_logprob: Optional[Union[List[bool], bool]] = False,