models:- model:NousResearch/Meta-Llama-3-8B# No parameters necessary for base model- model:NousResearch/Meta-Llama-3-8B-Instructparameters:density:0.6weight:0.55- model:mlabonne/OrpoLlama-3-8Bparameters:density:0.55weight:0.05- model:cognitivecomputations/dolphin-2.9-llama3-8bparameters:density:0.55weight:0.1- model:Locutusque/llama-3-neural-chat-v1-8bparameters:density:0.55weight:0.05- model:cloudyu/Meta-Llama-3-8B-Instruct-DPOparameters:density:0.55weight:0.15- model:vicgalle/Configurable-Llama-3-8B-v0.3parameters:density:0.55weight:0.1merge_method:dare_tiesbase_model:NousResearch/Meta-Llama-3-8Bparameters:int8_mask:truedtype:float16
💻 Usage
!pipinstall-qUtransformersacceleratefromtransformersimportAutoTokenizerimporttransformersimporttorchmodel="mlabonne/ChimeraLlama-3-8B-v2"messages=[{"role":"user","content":"What is a large language model?"}]tokenizer=AutoTokenizer.from_pretrained(model)prompt=tokenizer.apply_chat_template(messages,tokenize=False,add_generation_prompt=True)pipeline=transformers.pipeline("text-generation",model=model,torch_dtype=torch.float16,device_map="auto",)outputs=pipeline(prompt,max_new_tokens=256,do_sample=True,temperature=0.7,top_k=50,top_p=0.95)print(outputs[0]["generated_text"])