Co-authored-by: Ying Sheng <sqy1415@gmail.com> Co-authored-by: Liangsheng Yin <hnyls2002@gmail.com> Co-authored-by: Zhiqiang Xie <xiezhq@stanford.edu> Co-authored-by: parasol-aser <3848358+parasol-aser@users.noreply.github.com> Co-authored-by: LiviaSun <33578456+ChuyueSun@users.noreply.github.com> Co-authored-by: Cody Yu <hao.yu.cody@gmail.com>
127 lines
3.9 KiB
Python
127 lines
3.9 KiB
Python
import argparse
|
|
import asyncio
|
|
from concurrent.futures import ThreadPoolExecutor
|
|
from functools import partial
|
|
import json
|
|
import time
|
|
|
|
from tqdm import tqdm
|
|
import numpy as np
|
|
from sglang.test.test_utils import add_common_other_args_and_parse, call_generate_lightllm, call_generate_vllm, call_generate_srt_raw
|
|
from sglang.utils import read_jsonl, dump_state_text
|
|
|
|
|
|
USER_PREFIX = "[INST] "
|
|
USER_SUFFIX = " [/INST]"
|
|
ASSISTANT_PREFIX = ""
|
|
ASSISTANT_SUFFIX = " </s><s>"
|
|
|
|
|
|
def multi_document_qa(docs, question, generate):
|
|
s = USER_PREFIX
|
|
s += "Pleaes answer a question according to given documents.\n"
|
|
s += "Question:" + question + "Documents begin.\n"
|
|
|
|
s += "".join(docs)
|
|
|
|
s += "\nDocuments end."
|
|
s += ("\n\nBased on the above documents, please answer this question:\n" + question + "\nAnswer in three words or fewer.")
|
|
s += USER_SUFFIX
|
|
s += ASSISTANT_PREFIX
|
|
answer = generate(s, max_tokens=16, stop=None)
|
|
return answer
|
|
|
|
|
|
def main(args):
|
|
lines = read_jsonl(args.data_path)
|
|
l = lines[0]
|
|
arguments = []
|
|
labels = []
|
|
|
|
num_docs = 10
|
|
if args.backend == "guidance":
|
|
num_docs = 7 # due to OOM
|
|
|
|
for i in range(len(l["questions"][:args.num_questions])):
|
|
arguments.append({
|
|
"docs": l["documents"][:num_docs],
|
|
"question": l["questions"][i],
|
|
})
|
|
labels.append(l["answers"][i])
|
|
states = [None] * len(arguments)
|
|
|
|
# Select backend
|
|
if args.backend == "lightllm":
|
|
url = f"{args.host}:{args.port}/generate"
|
|
generate = partial(call_generate_lightllm, url=url, temperature=0)
|
|
elif args.backend == "vllm":
|
|
url = f"{args.host}:{args.port}/generate"
|
|
generate = partial(call_generate_vllm, url=url, temperature=0)
|
|
elif args.backend == "srt-raw":
|
|
url = f"{args.host}:{args.port}/generate"
|
|
generate = partial(call_generate_srt_raw, url=url, temperature=0)
|
|
elif args.backend == "guidance":
|
|
from guidance import models, gen
|
|
|
|
model = models.LlamaCpp("/home/ubuntu/model_weights/CodeLlama-7b-instruct-hf.gguf", n_gpu_layers=-1, n_ctx=11000)
|
|
|
|
def generate(prompt, max_tokens, stop):
|
|
out = model + prompt + gen(name="answer",
|
|
max_tokens=max_tokens, temperature=0, stop=stop)
|
|
return out["answer"]
|
|
|
|
# warmup
|
|
generate("Hello!", max_tokens=8, stop=None)
|
|
else:
|
|
raise ValueError(f"Invalid backend: {args.backend}")
|
|
|
|
# Run requests
|
|
def get_one_answer(i):
|
|
states[i] = multi_document_qa(generate=generate, **arguments[i])
|
|
|
|
tic = time.time()
|
|
if args.parallel == 1:
|
|
for i in tqdm(range(len(labels))):
|
|
get_one_answer(i)
|
|
else:
|
|
with ThreadPoolExecutor(args.parallel) as executor:
|
|
executor.map(get_one_answer, list(range(len(labels))))
|
|
latency = time.time() - tic
|
|
|
|
# Compute accuracy
|
|
print(states)
|
|
correct = 0
|
|
for s, label in zip(states, labels):
|
|
answer = s.lower()
|
|
if all(x in answer for x in label.lower().split(" ")):
|
|
correct += 1
|
|
accuracy = correct / len(labels)
|
|
print(f"Accuracy: {accuracy:.3f}")
|
|
print(f"Latency: {latency:.3f}")
|
|
|
|
# Write results
|
|
dump_state_text(f"tmp_output_{args.backend}.txt", states)
|
|
|
|
with open(args.result_file, "a") as fout:
|
|
value = {
|
|
"task": "multi_document_qa",
|
|
"backend": args.backend,
|
|
"num_gpus": 1,
|
|
"latency": round(latency, 3),
|
|
"num_requests": args.num_questions,
|
|
"accuracy": accuracy,
|
|
"other": {
|
|
"num_questions": args.num_questions,
|
|
"parallel": args.parallel,
|
|
}
|
|
}
|
|
fout.write(json.dumps(value) + "\n")
|
|
|
|
|
|
if __name__ == "__main__":
|
|
parser = argparse.ArgumentParser()
|
|
parser.add_argument("--data-path", type=str, default="questions.jsonl")
|
|
parser.add_argument("--num-questions", type=int, default=100)
|
|
args = add_common_other_args_and_parse(parser)
|
|
main(args)
|