192 lines
5.6 KiB
Markdown
192 lines
5.6 KiB
Markdown
|
|
---
|
|||
|
|
license: other
|
|||
|
|
license_name: qwen-research
|
|||
|
|
license_link: https://huggingface.co/Qwen/Qwen2.5-3B/blob/main/LICENSE
|
|||
|
|
library_name: transformers
|
|||
|
|
language:
|
|||
|
|
- ko
|
|||
|
|
- en
|
|||
|
|
tags:
|
|||
|
|
- text-generation
|
|||
|
|
- korean
|
|||
|
|
- bilingual
|
|||
|
|
- qwen2
|
|||
|
|
- built-with-qwen
|
|||
|
|
- continued-pretraining
|
|||
|
|
base_model: Qwen/Qwen2.5-3B
|
|||
|
|
datasets:
|
|||
|
|
- HuggingFaceFW/fineweb-edu
|
|||
|
|
- uonlp/CulturaX
|
|||
|
|
- wikimedia/wikipedia
|
|||
|
|
pipeline_tag: text-generation
|
|||
|
|
---
|
|||
|
|
|
|||
|
|
# 🐻 Gumini-1B (구미니)
|
|||
|
|
|
|||
|
|
<p align="center">
|
|||
|
|
<img src="https://img.shields.io/badge/Parameters-1.08B-blue" style="display:inline-block; margin-right:6px;" />
|
|||
|
|
<img src="https://img.shields.io/badge/Layers-10-green" style="display:inline-block; margin-right:6px;" />
|
|||
|
|
<img src="https://img.shields.io/badge/Languages-Korean%20%7C%20English-orange" style="display:inline-block; margin-right:6px;" />
|
|||
|
|
<img src="https://img.shields.io/badge/Built%20with-Qwen-purple" style="display:inline-block; margin-right:6px;" />
|
|||
|
|
</p>
|
|||
|
|
|
|||
|
|
<p align="center">
|
|||
|
|
<a href="https://linkedin.com/in/devgumin" target="_blank">
|
|||
|
|
<img src="https://img.shields.io/badge/LinkedIn-Gumin%20Kwon-0A66C2?logo=linkedin&logoColor=white" style="display:inline-block; margin-right:6px;" />
|
|||
|
|
</a>
|
|||
|
|
|
|||
|
|
<a href="https://x.com/Gumini_Research" target="_blank">
|
|||
|
|
<img src="https://img.shields.io/badge/X-@Gumini__Research-black?logo=x&logoColor=white" style="display:inline-block; margin-right:6px;" />
|
|||
|
|
</a>
|
|||
|
|
|
|||
|
|
<a href="https://www.instagram.com/gumini_research/" target="_blank">
|
|||
|
|
<img src="https://img.shields.io/badge/Instagram-gumini__research-E4405F?logo=instagram&logoColor=white" style="display:inline-block;" />
|
|||
|
|
</a>
|
|||
|
|
</p>
|
|||
|
|
|
|||
|
|
<p align="center"><b>Built with Qwen</b></p>
|
|||
|
|
|
|||
|
|
## Model Description
|
|||
|
|
|
|||
|
|
**Gumini** (구미니) is a bilingual Korean-English **base language model** created by inheriting the first 10 layers of **Qwen 2.5 3B** using the *Inheritune* methodology, followed by **continued pretraining** on a Korean–English mixed corpus (~393M tokens).
|
|||
|
|
|
|||
|
|
> This is a **BASE model**, not instruction-tuned.
|
|||
|
|
> It produces text continuations rather than conversational responses.
|
|||
|
|
|
|||
|
|
## What We Modified
|
|||
|
|
|
|||
|
|
The original **Qwen 2.5 3B** model was modified as follows:
|
|||
|
|
|
|||
|
|
1. **Layer Inheritance (Inheritune)**
|
|||
|
|
- Inherited the first **10 transformer layers** out of 36
|
|||
|
|
- Reduced model size while preserving early linguistic abilities
|
|||
|
|
|
|||
|
|
2. **Pretraining**
|
|||
|
|
- Trained for **393M tokens** on a Korean–English dataset
|
|||
|
|
- Maintains base-model behavior (not SFT or instruction-tuning)
|
|||
|
|
|
|||
|
|
3. **Identity Injection**
|
|||
|
|
- Added system-level identity tokens for model conditioning
|
|||
|
|
|
|||
|
|
This model **inherits early layers from Qwen 2.5 3B** and is
|
|||
|
|
**retrained with progressive layer expansion using Inheritune** methodology.
|
|||
|
|
|
|||
|
|
## Model Details
|
|||
|
|
|
|||
|
|
| Attribute | Value |
|
|||
|
|
|-----------|-------|
|
|||
|
|
| **Researcher** | Gumin Kwon (권구민) |
|
|||
|
|
| **Base Model** | Qwen/Qwen2.5-3B |
|
|||
|
|
| **Training Method** | Inheritune + Pretraining |
|
|||
|
|
| **Parameters** | 1.08B |
|
|||
|
|
| **Layers** | 10 |
|
|||
|
|
| **Hidden Size** | 2048 |
|
|||
|
|
| **Attention Heads** | 16 |
|
|||
|
|
| **KV Heads** | 2 (GQA) |
|
|||
|
|
| **Vocab Size** | 151,936 |
|
|||
|
|
| **Tokens Trained** | 393M |
|
|||
|
|
|
|||
|
|
## Training Data
|
|||
|
|
|
|||
|
|
| Dataset | Language | Weight |
|
|||
|
|
|---------|----------|---------|
|
|||
|
|
| FineWeb-Edu | English | 20% |
|
|||
|
|
| CulturaX-ko | Korean | 50% |
|
|||
|
|
| Wikipedia-ko | Korean | 30% |
|
|||
|
|
---
|
|||
|
|
## Usage
|
|||
|
|
|
|||
|
|
```python
|
|||
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|||
|
|
import torch
|
|||
|
|
|
|||
|
|
model = AutoModelForCausalLM.from_pretrained(
|
|||
|
|
"GuminiResearch/Gumini-1B-Base",
|
|||
|
|
torch_dtype=torch.bfloat16,
|
|||
|
|
device_map="auto"
|
|||
|
|
)
|
|||
|
|
tokenizer = AutoTokenizer.from_pretrained("GuminiResearch/Gumini-1B-Base")
|
|||
|
|
|
|||
|
|
prompt = "저는 구미니입니다."
|
|||
|
|
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
|||
|
|
outputs = model.generate(
|
|||
|
|
**inputs,
|
|||
|
|
max_new_tokens=100,
|
|||
|
|
repetition_penalty=1.2,
|
|||
|
|
do_sample=True,
|
|||
|
|
temperature=0.7
|
|||
|
|
)
|
|||
|
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
```python
|
|||
|
|
from transformers import pipeline
|
|||
|
|
|
|||
|
|
generator = pipeline(
|
|||
|
|
"text-generation",
|
|||
|
|
model="GuminiResearch/Gumini-1B-Base",
|
|||
|
|
)
|
|||
|
|
|
|||
|
|
prompt = "저는 구미니입니다."
|
|||
|
|
output = generator(prompt, max_new_tokens=100, temperature=0.7, repetition_penalty=1.2)
|
|||
|
|
|
|||
|
|
print(output[0]["generated_text"])
|
|||
|
|
```
|
|||
|
|
---
|
|||
|
|
## Limitations
|
|||
|
|
|
|||
|
|
- **Base model**: no instruction-tuning or safety alignment
|
|||
|
|
- **High repetition risk**: use `repetition_penalty >= 1.2`
|
|||
|
|
- May generate **incorrect or outdated information**
|
|||
|
|
- Should not be used in **sensitive or safety-critical** contexts
|
|||
|
|
|
|||
|
|
## License
|
|||
|
|
|
|||
|
|
### Qwen Research License (Non-Commercial)
|
|||
|
|
|
|||
|
|
This model is **Built with Qwen** and derived from Qwen 2.5 3B.
|
|||
|
|
|
|||
|
|
```
|
|||
|
|
Qwen is licensed under the Qwen RESEARCH LICENSE AGREEMENT.
|
|||
|
|
Copyright (c) Alibaba Cloud. All Rights Reserved.
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
**This model is for NON-COMMERCIAL / RESEARCH use only.**
|
|||
|
|
For commercial use, contact Alibaba Cloud.
|
|||
|
|
|
|||
|
|
### Inheritune Paper (CC BY 4.0)
|
|||
|
|
|
|||
|
|
```bibtex
|
|||
|
|
@inproceedings{Sanyal2024inheritune,
|
|||
|
|
title={Inheritune: Training Smaller Yet More Attentive Language Models},
|
|||
|
|
author={Sunny Sanyal and Ravid Shwartz-Ziv and Alexandros G. Dimakis and Sujay Sanghavi},
|
|||
|
|
year={2024},
|
|||
|
|
url={https://arxiv.org/abs/2404.08634}
|
|||
|
|
}
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
## Citation
|
|||
|
|
|
|||
|
|
```bibtex
|
|||
|
|
@misc{gumini2025,
|
|||
|
|
title={Gumini-1B: Bilingual Language Model Built with Qwen via Inheritune},
|
|||
|
|
author={Gumin Kwon},
|
|||
|
|
year={2025},
|
|||
|
|
note={Built with Qwen},
|
|||
|
|
url={https://huggingface.co/GuminiResearch/Gumini-1B-Base}
|
|||
|
|
}
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
## Author
|
|||
|
|
|
|||
|
|
**[Gumin Kwon (권구민)](https://linkedin.com/in/devgumin)**
|
|||
|
|
- LinkedIn: [linkedin.com/in/devgumin](https://linkedin.com/in/devgumin)
|
|||
|
|
- HuggingFace: [GuminiResearch](https://huggingface.co/GuminiResearch)
|
|||
|
|
|
|||
|
|
---
|
|||
|
|
|
|||
|
|
<p align="center">
|
|||
|
|
<b>Built with Qwen</b><br>
|
|||
|
|
<i>Gumini - 작지만 똑똑한 AI</i>
|
|||
|
|
</p>
|