diff --git a/README.md b/README.md index 4ebbf7b..cbd4102 100644 --- a/README.md +++ b/README.md @@ -83,10 +83,10 @@ model_path = "itpossible/Chinese-Mistral-7B-v0.1" tokenizer = AutoTokenizer.from_pretrained(model_path) model = AutoModelForCausalLM.from_pretrained(model_path, torch_dtype=torch.bfloat16, device_map=device) -text = "你是一个人工智能助手,写一段自我介绍:" +text = "我是一个人工智能助手,我能够帮助你做如下这些事情:" inputs = tokenizer(text, return_tensors="pt").to(device) -outputs = model.generate(**inputs, max_new_tokens=300) +outputs = model.generate(**inputs, max_new_tokens=120, do_sample=True) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` @@ -105,16 +105,16 @@ text = "请为我推荐中国三座比较著名的山" messages = [{"role": "user", "content": text}] inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(device) -outputs_id = model.generate(inputs, max_new_tokens=300, do_sample=True) -outputs = tokenizer.batch_decode(outputs_id, skip_special_tokens=True)[0] +outputs = model.generate(inputs, max_new_tokens=300, do_sample=True) +outputs = tokenizer.batch_decode(outputs, skip_special_tokens=True)[0] print(outputs) ``` Chinese-Mistral-7B 与Chinese-Mistral-7B-Instruct-v0.1生成效果示例如下: -| 模型名称 | Chinese-Mistral-7B | Chinese-Mistral-7B-Instruct-v0.1 | +| 模型名称 | Chinese-Mistral-7B | Chinese-Mistral-7B-Instruct | |--- |--- |--- | -| 用户输入 |