Upload README.md with huggingface_hub

This commit is contained in:
Bartowski
2025-05-05 16:24:11 +00:00
committed by system
parent f6df97c0bb
commit 534ac93b5a

View File

@@ -1,14 +1,11 @@
---
quantized_by: bartowski
pipeline_tag: text-generation
tags: []
base_model: mlabonne/Qwen3-14B-abliterated
base_model_relation: quantized
---
## Llamacpp imatrix Quantizations of Qwen3-14B-abliterated by mlabonne
Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b5200">b5200</a> for quantization.
Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b5270">b5270</a> for quantization.
Original model: https://huggingface.co/mlabonne/Qwen3-14B-abliterated
@@ -28,6 +25,10 @@ Run them directly with [llama.cpp](https://github.com/ggerganov/llama.cpp), or a
<|im_start|>assistant
```
## What's new:
New model files from mlabonne (WIP model)
## Download a file (not the whole branch) from below:
| Filename | Quant type | File Size | Split | Description |
@@ -56,6 +57,7 @@ Run them directly with [llama.cpp](https://github.com/ggerganov/llama.cpp), or a
| [Qwen3-14B-abliterated-IQ3_XXS.gguf](https://huggingface.co/bartowski/mlabonne_Qwen3-14B-abliterated-GGUF/blob/main/mlabonne_Qwen3-14B-abliterated-IQ3_XXS.gguf) | IQ3_XXS | 5.94GB | false | Lower quality, new method with decent performance, comparable to Q3 quants. |
| [Qwen3-14B-abliterated-Q2_K.gguf](https://huggingface.co/bartowski/mlabonne_Qwen3-14B-abliterated-GGUF/blob/main/mlabonne_Qwen3-14B-abliterated-Q2_K.gguf) | Q2_K | 5.75GB | false | Very low quality but surprisingly usable. |
| [Qwen3-14B-abliterated-IQ2_M.gguf](https://huggingface.co/bartowski/mlabonne_Qwen3-14B-abliterated-GGUF/blob/main/mlabonne_Qwen3-14B-abliterated-IQ2_M.gguf) | IQ2_M | 5.32GB | false | Relatively low quality, uses SOTA techniques to be surprisingly usable. |
| [Qwen3-14B-abliterated-IQ2_S.gguf](https://huggingface.co/bartowski/mlabonne_Qwen3-14B-abliterated-GGUF/blob/main/mlabonne_Qwen3-14B-abliterated-IQ2_S.gguf) | IQ2_S | 4.96GB | false | Low quality, uses SOTA techniques to be usable. |
## Embed/output weights