diff --git a/README.md b/README.md index cceff5e..73786e9 100644 --- a/README.md +++ b/README.md @@ -9,4 +9,30 @@ datasets: library_name: transformers --- -base Yi-6B-200k \ No newline at end of file +## Examples +```python +from transformers import AutoModelForCausalLM, AutoTokenizer + +model_path = 'chinoll/Yi-6b-200k-dpo' + +tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) + +# Since transformers 4.35.0, the GPT-Q/AWQ model can be loaded using AutoModelForCausalLM. +model = AutoModelForCausalLM.from_pretrained( + model_path, + device_map="auto", + torch_dtype='auto' +).eval() + +# Prompt content: "hi" +messages = [ + {"role": "user", "content": "hi"} +] + +input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt') +output_ids = model.generate(input_ids.to('cuda')) +response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True) + +# Model response: "Hello! How can I assist you today?" +print(response) +```