Files
Gumini-1B-Base/README.md

192 lines
5.6 KiB
Markdown
Raw Permalink Normal View History

---
license: other
license_name: qwen-research
license_link: https://huggingface.co/Qwen/Qwen2.5-3B/blob/main/LICENSE
library_name: transformers
language:
- ko
- en
tags:
- text-generation
- korean
- bilingual
- qwen2
- built-with-qwen
- continued-pretraining
base_model: Qwen/Qwen2.5-3B
datasets:
- HuggingFaceFW/fineweb-edu
- uonlp/CulturaX
- wikimedia/wikipedia
pipeline_tag: text-generation
---
# 🐻 Gumini-1B (구미니)
<p align="center">
<img src="https://img.shields.io/badge/Parameters-1.08B-blue" style="display:inline-block; margin-right:6px;" />
<img src="https://img.shields.io/badge/Layers-10-green" style="display:inline-block; margin-right:6px;" />
<img src="https://img.shields.io/badge/Languages-Korean%20%7C%20English-orange" style="display:inline-block; margin-right:6px;" />
<img src="https://img.shields.io/badge/Built%20with-Qwen-purple" style="display:inline-block; margin-right:6px;" />
</p>
<p align="center">
<a href="https://linkedin.com/in/devgumin" target="_blank">
<img src="https://img.shields.io/badge/LinkedIn-Gumin%20Kwon-0A66C2?logo=linkedin&logoColor=white" style="display:inline-block; margin-right:6px;" />
</a>
<a href="https://x.com/Gumini_Research" target="_blank">
<img src="https://img.shields.io/badge/X-@Gumini__Research-black?logo=x&logoColor=white" style="display:inline-block; margin-right:6px;" />
</a>
<a href="https://www.instagram.com/gumini_research/" target="_blank">
<img src="https://img.shields.io/badge/Instagram-gumini__research-E4405F?logo=instagram&logoColor=white" style="display:inline-block;" />
</a>
</p>
<p align="center"><b>Built with Qwen</b></p>
## Model Description
**Gumini** (구미니) is a bilingual Korean-English **base language model** created by inheriting the first 10 layers of **Qwen 2.5 3B** using the *Inheritune* methodology, followed by **continued pretraining** on a KoreanEnglish mixed corpus (~393M tokens).
> This is a **BASE model**, not instruction-tuned.
> It produces text continuations rather than conversational responses.
## What We Modified
The original **Qwen 2.5 3B** model was modified as follows:
1. **Layer Inheritance (Inheritune)**
- Inherited the first **10 transformer layers** out of 36
- Reduced model size while preserving early linguistic abilities
2. **Pretraining**
- Trained for **393M tokens** on a KoreanEnglish dataset
- Maintains base-model behavior (not SFT or instruction-tuning)
3. **Identity Injection**
- Added system-level identity tokens for model conditioning
This model **inherits early layers from Qwen 2.5 3B** and is
**retrained with progressive layer expansion using Inheritune** methodology.
## Model Details
| Attribute | Value |
|-----------|-------|
| **Researcher** | Gumin Kwon (권구민) |
| **Base Model** | Qwen/Qwen2.5-3B |
| **Training Method** | Inheritune + Pretraining |
| **Parameters** | 1.08B |
| **Layers** | 10 |
| **Hidden Size** | 2048 |
| **Attention Heads** | 16 |
| **KV Heads** | 2 (GQA) |
| **Vocab Size** | 151,936 |
| **Tokens Trained** | 393M |
## Training Data
| Dataset | Language | Weight |
|---------|----------|---------|
| FineWeb-Edu | English | 20% |
| CulturaX-ko | Korean | 50% |
| Wikipedia-ko | Korean | 30% |
---
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model = AutoModelForCausalLM.from_pretrained(
"GuminiResearch/Gumini-1B-Base",
torch_dtype=torch.bfloat16,
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("GuminiResearch/Gumini-1B-Base")
prompt = "저는 구미니입니다."
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(
**inputs,
max_new_tokens=100,
repetition_penalty=1.2,
do_sample=True,
temperature=0.7
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
```python
from transformers import pipeline
generator = pipeline(
"text-generation",
model="GuminiResearch/Gumini-1B-Base",
)
prompt = "저는 구미니입니다."
output = generator(prompt, max_new_tokens=100, temperature=0.7, repetition_penalty=1.2)
print(output[0]["generated_text"])
```
---
## Limitations
- **Base model**: no instruction-tuning or safety alignment
- **High repetition risk**: use `repetition_penalty >= 1.2`
- May generate **incorrect or outdated information**
- Should not be used in **sensitive or safety-critical** contexts
## License
### Qwen Research License (Non-Commercial)
This model is **Built with Qwen** and derived from Qwen 2.5 3B.
```
Qwen is licensed under the Qwen RESEARCH LICENSE AGREEMENT.
Copyright (c) Alibaba Cloud. All Rights Reserved.
```
**This model is for NON-COMMERCIAL / RESEARCH use only.**
For commercial use, contact Alibaba Cloud.
### Inheritune Paper (CC BY 4.0)
```bibtex
@inproceedings{Sanyal2024inheritune,
title={Inheritune: Training Smaller Yet More Attentive Language Models},
author={Sunny Sanyal and Ravid Shwartz-Ziv and Alexandros G. Dimakis and Sujay Sanghavi},
year={2024},
url={https://arxiv.org/abs/2404.08634}
}
```
## Citation
```bibtex
@misc{gumini2025,
title={Gumini-1B: Bilingual Language Model Built with Qwen via Inheritune},
author={Gumin Kwon},
year={2025},
note={Built with Qwen},
url={https://huggingface.co/GuminiResearch/Gumini-1B-Base}
}
```
## Author
**[Gumin Kwon (권구민)](https://linkedin.com/in/devgumin)**
- LinkedIn: [linkedin.com/in/devgumin](https://linkedin.com/in/devgumin)
- HuggingFace: [GuminiResearch](https://huggingface.co/GuminiResearch)
---
<p align="center">
<b>Built with Qwen</b><br>
<i>Gumini - 작지만 똑똑한 AI</i>
</p>