Upload README.md with huggingface_hub

This commit is contained in:
Bartowski
2025-03-12 07:40:42 +00:00
committed by system
parent a619a1e1a6
commit 7465f519b9

View File

@@ -1,8 +1,6 @@
---
quantized_by: bartowski
pipeline_tag: text-generation
license: apache-2.0
base_model: RekaAI/reka-flash-3
---
## Llamacpp imatrix Quantizations of reka-flash-3 by RekaAI
@@ -20,9 +18,13 @@ Run them directly with [llama.cpp](https://github.com/ggerganov/llama.cpp), or a
## Prompt format
```
human: {prompt} <sep> assistant:
human: {system_prompt} {prompt} <sep> assistant:
```
## What's new:
Fix chat template
## Download a file (not the whole branch) from below:
| Filename | Quant type | File Size | Split | Description |
@@ -52,6 +54,8 @@ human: {prompt} <sep> assistant:
| [reka-flash-3-Q2_K.gguf](https://huggingface.co/bartowski/RekaAI_reka-flash-3-GGUF/blob/main/RekaAI_reka-flash-3-Q2_K.gguf) | Q2_K | 8.63GB | false | Very low quality but surprisingly usable. |
| [reka-flash-3-IQ2_M.gguf](https://huggingface.co/bartowski/RekaAI_reka-flash-3-GGUF/blob/main/RekaAI_reka-flash-3-IQ2_M.gguf) | IQ2_M | 8.51GB | false | Relatively low quality, uses SOTA techniques to be surprisingly usable. |
| [reka-flash-3-IQ2_S.gguf](https://huggingface.co/bartowski/RekaAI_reka-flash-3-GGUF/blob/main/RekaAI_reka-flash-3-IQ2_S.gguf) | IQ2_S | 8.12GB | false | Low quality, uses SOTA techniques to be usable. |
| [reka-flash-3-IQ2_XS.gguf](https://huggingface.co/bartowski/RekaAI_reka-flash-3-GGUF/blob/main/RekaAI_reka-flash-3-IQ2_XS.gguf) | IQ2_XS | 7.83GB | false | Low quality, uses SOTA techniques to be usable. |
| [reka-flash-3-IQ2_XXS.gguf](https://huggingface.co/bartowski/RekaAI_reka-flash-3-GGUF/blob/main/RekaAI_reka-flash-3-IQ2_XXS.gguf) | IQ2_XXS | 7.39GB | false | Very low quality, uses SOTA techniques to be usable. |
## Embed/output weights