56 lines
2.3 KiB
Markdown
56 lines
2.3 KiB
Markdown
|
|
---
|
||
|
|
license: apache-2.0
|
||
|
|
language:
|
||
|
|
- en
|
||
|
|
- de
|
||
|
|
- es
|
||
|
|
- fr
|
||
|
|
- it
|
||
|
|
base_model:
|
||
|
|
- VillanovaAI/Villanova-2B-2603
|
||
|
|
pipeline_tag: text-generation
|
||
|
|
---
|
||
|
|
|
||
|
|
|
||
|
|
# Model Card for Villanova-2B-2603-GGUF
|
||
|
|
|
||
|
|
<img src="https://huggingface.co/spaces/VillanovaAI/README/resolve/main/Logo_VILLANOVA_colore.svg" alt="Villanova.AI logo" height="96"/>
|
||
|
|
|
||
|
|
**Villanova-2B-2603** is a fully open, multilingual instruction-tuned Large Language Model developed by [Villanova.AI](https://huggingface.co/VillanovaAI). Part of the Villanova project, it is designed to advance open European language technology with native support for five European languages. All model weights, training data sources, and training details are publicly released.
|
||
|
|
|
||
|
|
This repo contains GGUF format model files for the [VillanovaAI/Villanova-2B-2603](https://huggingface.co/VillanovaAI/Villanova-2B-2603) model.
|
||
|
|
|
||
|
|
---
|
||
|
|
|
||
|
|
## Model Family
|
||
|
|
|
||
|
|
**[Villanova-2B-Base-2603](https://huggingface.co/VillanovaAI/Villanova-2B-Base-2603)** — Base model (4.4T)<br>
|
||
|
|
 ↳ **[Villanova-2B-2603](https://huggingface.co/VillanovaAI/Villanova-2B-2603)** — SFT / Instruct<br>
|
||
|
|
  ↳ [Villanova-2B-2603-GGUF](https://huggingface.co/VillanovaAI/Villanova-2B-2603-GGUF) — Quantized — 📍 *This model*<br>
|
||
|
|
 ↳ **[Villanova-2B-VL-2603](https://huggingface.co/VillanovaAI/Villanova-2B-VL-2603)** — Vision-Language Instruct<br>
|
||
|
|
  ↳ [Villanova-2B-VL-2603-GGUF](https://huggingface.co/VillanovaAI/Villanova-2B-VL-2603-GGUF) — Quantized<br>
|
||
|
|
<br>
|
||
|
|
**[Villanova-2B-Base-2512-Preview](https://huggingface.co/VillanovaAI/Villanova-2B-Base-2512-Preview)** — Base model (2.2T) (previous version, not recommended)<br>
|
||
|
|
 ↳ **[Villanova-2B-2512-Preview](https://huggingface.co/VillanovaAI/Villanova-2B-2512-Preview)** — SFT / Instruct (previous version, not recommended)<br>
|
||
|
|
|
||
|
|
|
||
|
|
|
||
|
|
|
||
|
|
## About GGUF
|
||
|
|
**GGUF** is a format introduced by llama.cpp.
|
||
|
|
|
||
|
|
It is a file format for storing and distributing LLMs that is designed for portability and efficient **inference on the edge**.
|
||
|
|
|
||
|
|
## Quick Usage with llama.cpp
|
||
|
|
|
||
|
|
You can run this model directly using the `llama-cli` tool (part of [llama.cpp](https://github.com/ggerganov/llama.cpp)).
|
||
|
|
|
||
|
|
To run the model with the **Q8_0** quantization directly from Hugging Face:
|
||
|
|
|
||
|
|
```bash
|
||
|
|
llama-cli -hf VillanovaAI/Villanova-2B-2603-GGUF:Q8_0
|
||
|
|
```
|
||
|
|
|
||
|
|
|
||
|
|
|