Files
Llama3.2-HermesDolphin-Code…/README.md

70 lines
2.0 KiB
Markdown
Raw Normal View History

---
license: other
base_model:
- artificialguybr/LLAMA-3.2-1B-OpenHermes2.5
- dphn/Dolphin3.0-Llama3.2-1B
- meta-llama/Llama-3.2-1B-Instruct
library_name: transformers
pipeline_tag: text-generation
tags:
- llama
- mergekit
- merge
- slerp
- text-generation
- code
- instruct
datasets:
- OpenCoder-LLM/opc-sft-stage1
- OpenCoder-LLM/opc-sft-stage2
- microsoft/orca-agentinstruct-1M-v1
- microsoft/orca-math-word-problems-200k
- NousResearch/hermes-function-calling-v1
- AI-MO/NuminaMath-CoT
- AI-MO/NuminaMath-TIR
- allenai/tulu-3-sft-mixture
- HuggingFaceTB/smoltalk
- m-a-p/CodeFeedback-Filtered-Instruction
- m-a-p/Code-Feedback
- teknium/OpenHermes-2.5
---
# Llama-3.2-HermesDolphin-Coder-1B
Llama-3.2-HermesDolphin-Coder-1B is a compact merged language model designed for general instruction following, coding assistance, and lightweight conversational use. It combines Hermes-style instruction tuning and Dolphin-style helpfulness into a small Llama 3.2 class model intended for experimentation, local workflows, and developer-oriented prompting.
This repository appears to be a **merge model** created with **mergekit** using the **SLERP** merge method.
## Model Summary
- **Model type:** Causal language model
- **Architecture:** LlamaForCausalLM
- **Primary use:** Text generation, instruction following, code-oriented prompting
- **Library:** Transformers
- **Merge method:** SLERP
- **Format:** Safetensors
## Base Models
This merged model is based on:
- `artificialguybr/LLAMA-3.2-1B-OpenHermes2.5`
- `dphn/Dolphin3.0-Llama3.2-1B`
- `meta-llama/Llama-3.2-1B-Instruct`
## Merge Details
According to the repository metadata/configuration, the merge was produced with `mergekit` using a SLERP setup with a midpoint interpolation parameter.
### Merge configuration
```yaml
merge_method: slerp
base_model: artificialguybr/LLAMA-3.2-1B-OpenHermes2.5
models:
- model: dphn/Dolphin3.0-Llama3.2-1B
parameters:
weight: 1.0
dtype: float32
parameters:
t: 0.5