初始化项目,由ModelHub XC社区提供模型

Model: tartuNLP/Llama-2-7b-Ukrainian
Source: Original Platform
This commit is contained in:
ModelHub XC
2026-04-22 00:43:36 +08:00
commit 4680e50981
11 changed files with 500 additions and 0 deletions

50
README.md Normal file
View File

@@ -0,0 +1,50 @@
---
library_name: transformers
license: llama2
datasets:
- uonlp/CulturaX
language:
- uk
- en
pipeline_tag: text-generation
base_model:
- meta-llama/Llama-2-7b-hf
---
# Llama-2-7b-Ukrainian
## Model Details
### Model Description
Llama-2-7b-Ukrainian is a bilingual pre-trained model supporting Ukrainian and English.
Continued pre-training from [Llama-2-7b](https://huggingface.co/meta-llama/Llama-2-7b-hf) on 5B tokens
consisting of 75% Ukrainian documents and 25% English documents from [CulturaX](https://huggingface.co/datasets/uonlp/CulturaX).
**Paper:** [To Err Is Human, but Llamas Can Learn It Too](https://arxiv.org/abs/2403.05493)
### Training Hyperparameters
| Hyperparameter | Value |
|---|---|
| Training steps | 19080 |
| Batch size | 256 |
| Weight decay | 0.1 |
| Context length | 1024 |
| Learning rate | 2e-5 linear decay to 2e-6 |
| Precision | bf16 |
| Optimizer | AdamW |
## Citation
**BibTeX:**
```
@article{luhtaru2024err,
title={To Err Is Human, but Llamas Can Learn It Too},
author={Luhtaru, Agnes and Purason, Taido and Vainikko, Martin and Del, Maksym and Fishel, Mark},
journal={arXiv preprint arXiv:2403.05493},
year={2024}
}
```