Return logprob for choices (#87)
This commit is contained in:
36
examples/usage/async_io.py
Normal file
36
examples/usage/async_io.py
Normal file
@@ -0,0 +1,36 @@
|
||||
import asyncio
|
||||
from sglang import Runtime
|
||||
|
||||
|
||||
async def generate(
|
||||
engine,
|
||||
prompt,
|
||||
sampling_params,
|
||||
):
|
||||
tokenizer = engine.get_tokenizer()
|
||||
|
||||
messages = [
|
||||
{"role": "system", "content": "You will be given question answer tasks.",},
|
||||
{"role": "user", "content": prompt},
|
||||
]
|
||||
|
||||
prompt = tokenizer.apply_chat_template(
|
||||
messages, tokenize=False, add_generation_prompt=True
|
||||
)
|
||||
|
||||
stream = engine.add_request(prompt, sampling_params)
|
||||
|
||||
async for output in stream:
|
||||
print(output, end="", flush=True)
|
||||
print()
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
runtime = Runtime(model_path="meta-llama/Llama-2-7b-chat-hf")
|
||||
print("runtime ready")
|
||||
|
||||
prompt = "Who is Alan Turing?"
|
||||
sampling_params = {"max_new_tokens": 128}
|
||||
asyncio.run(generate(runtime, prompt, sampling_params))
|
||||
|
||||
runtime.shutdown()
|
||||
Reference in New Issue
Block a user