Files
Aitana-7B-S-base-1.0/README.md
ModelHub XC 5ed70308a7 初始化项目,由ModelHub XC社区提供模型
Model: gplsi/Aitana-7B-S-base-1.0
Source: Original Platform
2026-04-23 20:14:07 +08:00

305 lines
18 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

---
license: apache-2.0
language:
- ca
- es
- en
base_model: BSC-LT/salamandra-7b
tags:
- valencian
- catalan
- spanish
- english
- text-generation
- alia
- gplsi
datasets:
- gplsi/alia_dogv
- gplsi/alia_les_corts
- gplsi/alia_amic
- gplsi/alia_boua
- gplsi/alia_tourism
library_name: transformers
pipeline_tag: text-generation
---
# Aitana-7B-S-base-1.0
**Aitana-7B-S-base-1.0** is a generative language model from the **Aitana family**, developed by the [GPLSI (Language and Information System Group)](https://gplsi.dlsi.ua.es/) at the University of Alicante. This model is based on [BSC-LT/salamandra-7b](https://huggingface.co/BSC-LT/salamandra-7b) and has been continuously pre-trained on multilingual data (Valencian, Spanish, and English) to improve representation of Valencian and Catalan languages.
## Table of Contents
- [Model Description](#model-description)
- [Evaluation](#evaluation)
- [Training Data](#training-data)
- [Intended Uses](#intended-uses)
- [How to Use](#how-to-use)
- [Additional Information](#additional-information)
## Model Description
| Property | Value |
|----------|-------|
| **Base Model** | [BSC-LT/salamandra-7b](https://huggingface.co/BSC-LT/salamandra-7b) |
| **Architecture** | Transformer decoder-only |
| **Parameters** | ~7.77B |
| **Languages** | Valencian, Spanish, English |
| **License** | Apache 2.0 |
Aitana-7B-S-base-1.0 extends the multilingual Salamandra foundation with additional training on domain-specific Valencian, Spanish, and English data. The training emphasizes administrative, legal, and tourism domains.
## Training Data
This model was trained on the following ALIA datasets:
| Dataset ID | Name | Language | Source |
|------------|------|----------|--------|
| dc8 | dogv_va_2025 | Valencian | [gplsi/alia_dogv](https://huggingface.co/datasets/gplsi/alia_dogv) |
| dc9 | dogv_es_2025 | Spanish | [gplsi/alia_dogv](https://huggingface.co/datasets/gplsi/alia_dogv) |
| dc10 | corts_es_va_2025 | Spanish/Valencian | [gplsi/alia_les_corts](https://huggingface.co/datasets/gplsi/alia_les_corts) |
| dc11 | amic_va_2025 | Valencian | [gplsi/alia_amic](https://huggingface.co/datasets/gplsi/alia_amic) |
| dc12 | boua_va_2025 | Valencian | [gplsi/alia_boua](https://huggingface.co/datasets/gplsi/alia_boua) |
| dc13 | boua_es_2025 | Spanish | [gplsi/alia_boua](https://huggingface.co/datasets/gplsi/alia_boua) |
| dc14 | tourism_va_2025 | Valencian | [gplsi/alia_tourism](https://huggingface.co/datasets/gplsi/alia_tourism) |
| dc15 | tourism_es_2025 | Spanish | [gplsi/alia_tourism](https://huggingface.co/datasets/gplsi/alia_tourism) |
| dc16 | tourism_en_2025 | English | [gplsi/alia_tourism](https://huggingface.co/datasets/gplsi/alia_tourism) |
### Data Sources
- **DOGV (Diari Oficial de la Generalitat Valenciana)**: Official communications of the Valencian Community including laws and public sector communications
- **Les Corts Valencianes**: Transcripts from the Valencian Parliament plenary sessions and committee meetings
- **AMIC**: Valencian language corpus
- **BOUA (Butlletí Oficial de la Universitat d'Alacant)**: Official University of Alicante documents including grants, regulations, and resolutions
- **Tourism**: Multilingual tourism domain content
## Intended Uses
This model can be used for:
- **Text generation** in Valencian, Spanish, and English
- **Fine-tuning** for specific downstream tasks
- **Domain adaptation** for administrative, legal, or tourism applications
> **Note**: Due to the formal register of training data (administrative and legal domains), generated text tends toward formal language.
## How to Use
### Transformers
```python
import torch
from transformers import pipeline, AutoTokenizer
model_id = "gplsi/Aitana-7B-S-base-1.0"
tokenizer = AutoTokenizer.from_pretrained(model_id)
generator = pipeline(
"text-generation",
model=model_id,
tokenizer=tokenizer,
torch_dtype=torch.bfloat16,
device_map="auto",
)
# Valencian example
text = "Les corts valencianes han pres la decisió de"
result = generator(text, do_sample=True, top_k=10, max_new_tokens=100)
print(result[0]['generated_text'])
# Spanish example
text = "El turismo en la Comunidad Valenciana"
result = generator(text, do_sample=True, top_k=10, max_new_tokens=100)
print(result[0]['generated_text'])
```
## Evaluation
In the following table, we can see the results obtained with different benchmarks from [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness) in comparison with the model used for continuous pre-training.
The results have been obtained from the model pre-trained; no instruction tuning or fine-tuning of any kind has been performed.
### Normalized score per language
| Language | Salamandra-7B | Aitana-7B-S-base-1.0 |
|----------|----------|----------|
| **Spanish** | **0.255** | 0.252 |
| **Catalan** | 0.373 | **0.378** |
| **English** | 0.329 | **0.364** |
| **Valencian** | **0.614** | **0.614** |
### Valencian
#### Classification Benchmarks
| Dataset | Lang. | Task | Metric | Salamandra-7B | Aitana-7B-S-base-1.0 |
|------------------------------|--------|----------------------------|-------------|---------------|-----------------------|
| XNLI | va |Natural Language Inference | acc | **0.50** | 0.50 |
#### Generation Benchmarks
| Dataset | Lang. | Task | Metric | Salamandra-7B | Aitana-7B-S-base-1.0 |
|------------------------------|--------|----------------------------|-------------|---------------|-----------------------|
| Cocoteros | va |Reading Comprehension | bleu | 12.01 | **16.19** |
| Phrases ca-va | va-ca |Translation - Adaptation | bleu | **86.80** | 85.33 |
| Phrases va-ca | va-ca |Translation - Adaptation | bleu | **94.71** | 80.00 |
| Phrases va-es | va-es |Translation | bleu | 79.74 | **80.59** |
| Phrases es-va | es-va |Translation | bleu | 66.42 | **69.78** |
| Truthfulqa_va | va | Truthfulness | bleu_acc| 0.33 | **0.37** |
### Catalan
#### Classification Benchmarks
| Dataset | Lang. | Task | Metric | Salamandra-7B | Aitana-7B-S-base-1.0 |
|------------------------------|--------|----------------------------|-------------|---------------|-----------------------|
| Belebele Cat_latn | ca | Reading Comprehension | acc | 0.51 | **0.54** |
| COPA | ca | Commonsense Reasoning | acc | 0.80 | **0.82** |
| XStoryCloze | ca | Commonsense Reasoning | acc | 0.75 | **0.77** |
| OpenBookQA | ca | Question Answering | acc | 0.38 | **0.38** |
| PAWS | ca | Paraphrasing | acc | 0.62 | **0.62** |
| PiQA | ca | Question Answering | acc | 0.71 | **0.72** |
| SiQA | ca | Question Answering | acc | 0.49 | **0.51** |
| ARC Easy | ca | Question Answering | acc | 0.73 | **0.73** |
| ARC Challenge | ca | Question Answering | acc | **0.47** | 0.46 |
| XNLI | ca | Natural Language Inference | acc | **0.51** | 0.50 |
| Teca | ca | Natural Language Inference | acc | 0.53 | **0.53** |
| WNLI | ca | Natural Language Inference | acc | 0.59 | **0.62** |
| Catcola | ca | Linguistic Acceptability | acc | **0.73** | 0.73 |
| Catcola | ca | Linguistic Acceptability | mcc | **0.29** | 0.15 |
| Catalanqa | ca | Question Answering | F1 | 0.82 | **0.83** |
| Mgsm direct | ca | Math | exact match | 0.07 | **0.09** |
| Catalanqa | ca | Question Answering | exact match | 0.62 | **0.65** |
| Xquad | ca | Question Answering | exact match | 0.49 | **0.51** |
| Xquad | ca | Question Answering | F1 | 0.71 | **0.73** |
#### Generation Benchmarks
| Dataset | Lang. | Task | Metric | Salamandra-7B | Aitana-7B-S-base-1.0 |
|------------------------------|--------|----------------------------|--------|----------------|-----------------------|
| Cabreu abstractive | ca | Summarization | bleu | 8.73 | **11.32** |
| Cabreu extractive | ca | Summarization | bleu | **44.55** | 41.80 |
| Cabreu extreme | ca | Summarization | bleu | 10.66 | **12.54** |
### Spanish
#### Classification Benchmarks
| Dataset | Lang. | Task | Metric | Salamandra-7B | Aitana-7B-S-base-1.0 |
|------------------------------|--------|---------------------------|-------------|---------------|-----------------------|
| Belebele | es | Reading Comprehension | acc | 0.493 | **0.561** |
| PAWS | es | Paraphrasing | acc | **0.608** | 0.591 |
| XNLI | es | Natural Language Inference| acc | **0.468** | 0.462 |
| WNLI | es | Natural Language Inference| acc | **0.465** | 0.437 |
| XStoryCloze | es | Commonsense Reasoning | acc | 0.745 | **0.756** |
| Escola | es | Linguistic Acceptability | acc | **0.706** | 0.678 |
| Escola | es | Linguistic Acceptability | mcc | **0.295** | 0.146 |
| OpenbookQA | es | Question Answering | acc | **0.406** | 0.382 |
| MGSM Direct | es | Math | exact match | 0.068 | **0.080** |
| XQUAD | es | Question Answering | exact match | 0.501 | **0.505** |
| XQUAD | es | Question Answering | F1 | 0.711 | **0.719** |
#### Generation Benchmarks
| Dataset | Lang. | Task | Metric | Salamandra-7B | Aitana-7B-S-base-1.0 |
|------------------------------|--------|---------------------|---------|----------------|-----------------------|
| Cocoteros | es |Reading Comprehension| bleu | 13.68 | **17.51** |
| XLSum | es | Summarization | bleu | 3.59 | **5.75** |
### English
#### Classification Benchmarks
| Dataset | Lang. | Task | Metric | Salamandra-7B | Aitana-7B-S-base-1.0 |
|------------------------------|--------|----------------------------|-------------|---------------|-----------------------|
| Arc Challenge | en | Question Answering | acc | **0.527** | 0.526 |
| Arc Easy | en | Question Answering | acc | **0.824** | 0.814 |
| Belebele | en | Reading Comprehension | acc | 0.549 | **0.573** |
| PAWS | en | Paraphrasing | acc | **0.633** | 0.615 |
| XNLI | en | Natural Language Inference | acc | **0.483** | 0.476 |
| XStoryCloze | en | Commonsense Reasoning | acc | **0.795** | 0.793 |
| OpenBookQA | en | Question Answering | acc | 0.356 | **0.362** |
| PiQA | en | Question Answering | acc | 0.797 | **0.799** |
| Social iqa | en | Question Answering | acc | **0.513** | 0.512 |
| WNLI | en | Natural Language Inference | acc | 0.479 | **0.606** |
| MGSM Direct | en | Math | exact match | 0.280 | **0.564** |
| TriviaQA | en | Question Answering | exact match | 0.597 | **0.602** |
| CoLA | en | Linguistic Acceptability | mcc | **0.412** | 0.361 |
## Additional Information
### Author
The model has been developed by the **Language and [Information Systems Group (GPLSI)](https://gplsi.dlsi.ua.es/)** and the **[Centro de Inteligencia Digital (CENID)](https://cenid.es)**, both part of the **[University of Alicante (UA)](https://www.ua.es/es/)**, as part of their ongoing research in **Natural Language Processing (NLP)**.
### Part of the Aitana Family
This model is part of the Aitana model family developed by the GPLSI research group, which includes:
- [gplsi/Aitana-2B-S](https://huggingface.co/gplsi/Aitana-2B-S) - Valencian-focused 2B model
- [gplsi/Aitana-2B-S-base-1.0](https://huggingface.co/gplsi/Aitana-2B-S-base-1.0) - Base version (1.0) of the 2B model
- [gplsi/Aitana-6.3B](https://huggingface.co/gplsi/Aitana-6.3B) - Larger 6.3B parameter model
- [gplsi/Aitana-TA-2B-S](https://huggingface.co/gplsi/Aitana-TA-2B-S) - Translation model (Spanish ↔ Valencian)
- [gplsi/Aitana-2B-S-LF](https://www.google.com/search?q=https://huggingface.co/gplsi/Aitana-2B-S-LF) - 2B Text Generation variant
- [gplsi/Aitana-2B-S-tourism-base-1.0](https://huggingface.co/gplsi/Aitana-2B-S-tourism-base-1.0) - Domain-specific base model focused on Tourism
- [gplsi/Aitana-tourism-mb-encoder-1.0](https://huggingface.co/gplsi/Aitana-tourism-mb-encoder-1.0) - Tourism domain Fill-Mask/Encoder model
- [gplsi/Aitana-FraudDetection-R-1.0](https://huggingface.co/gplsi/Aitana-FraudDetection-R-1.0) - Text Classification model for Fraud Detection
### Funding
This work is funded by the **Ministerio para la Transformación Digital y de la Función Pública**, co-financed by the **EU NextGenerationEU**, within the framework of the project *Desarrollo de Modelos ALIA*.
### Acknowledgments
We would like to express our gratitude to all individuals and institutions that have contributed to the development of this work.
Special thanks to:
- [Language Technologies Laboratory at Barcelona Supercomputing Center](https://www.bsc.es/es/discover-bsc/organisation/research-structure/language-technologies-laboratory)
- [Centro Vasco de Tecnología de la Lengua (HiTZ)](https://www.hitz.eus/es)
- [Centro Singular de Investigación en Tecnologías Inteligentes (CiTIUS)](https://citius.gal/)
- [Sistemas Inteligentes de Acceso a la Información (SINAI)](https://www.ujaen.es/investigacion-y-transferencia/grupos-de-investigacion/sistemas-inteligentes-de-acceso-la-informacion-sinai)
- [Instituto Universitario de Investigación Informática (IUII)](https://web.ua.es/es/iuii/)
- [Leonardo HPC System](https://leonardo-supercomputer.cineca.eu/)
- [European supercomputing ecosystem (EUROHPC)](https://www.eurohpc-ju.europa.eu/)
We also acknowledge the financial, technical, and scientific support of the **Ministerio para la Transformación Digital y de la Función Pública - Funded by EU NextGenerationEU within the framework of the project Desarrollo de Modelos ALIA**, whose contribution has been essential to the completion of this research.
### License
[Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0)
### Disclaimer
This model is intended for general purposes and is available under a permissive Apache License 2.0. Be aware that the model may have biases and/or undesirable outputs. Users deploying systems based on this model are responsible for mitigating risks and complying with applicable AI regulations.
### Reference
```bibtex
@misc{gplsi-aitana-2B-S-base-1.0,
author = {Estevanell-Valladares, Ernesto L. and Yáñez-Romero, Fabio and Sepúlveda-Torres, Robiert and Consuegra-Ayala, Juan Pablo and Galiano, Santiago and Miró Maestre, María and Martínez-Murillo, Iván and Grande, Eduardo and Canal-Esteve, Miquel and Bonora, Mar and Gutierrez, Yoan and Abreu Salas, José Ignacio and Lloret, Elena and Montoyo, Andrés and Muñoz-Guillena and Palomar, Manuel},
title = {Aitana 7B base: Continually pre-trained on Valencian},
year = {2026},
institution = {Language and Information Systems Group (GPLSI) and Centro de Inteligencia Digital (CENID), University of Alicante (UA)},
howpublished = {\url{https://huggingface.co/gplsi/gplsi/Aitana-2B-S-base-1.0}},
note = {Accessed: 2026-4-8}
}
```
---
**Copyright © 2026 Language and Information Systems Group (GPLSI) and Centro de Inteligencia Digital (CENID),
University of Alicante (UA).
Distributed under the Apache License 2.0.**