From 3543032c486eb93c1b975127a6dece3979ed4905 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sat, 9 May 2026 07:17:58 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/gemma-3-4b-it-qat-abliterated-GGUF Source: Original Platform --- .gitattributes | 47 +++++++++++++++++ README.md | 64 +++++++++++++++++++++++ gemma-3-4b-it-qat-abliterated.IQ4_XS.gguf | 3 ++ gemma-3-4b-it-qat-abliterated.Q2_K.gguf | 3 ++ gemma-3-4b-it-qat-abliterated.Q3_K_L.gguf | 3 ++ gemma-3-4b-it-qat-abliterated.Q3_K_M.gguf | 3 ++ gemma-3-4b-it-qat-abliterated.Q3_K_S.gguf | 3 ++ gemma-3-4b-it-qat-abliterated.Q4_K_M.gguf | 3 ++ gemma-3-4b-it-qat-abliterated.Q4_K_S.gguf | 3 ++ gemma-3-4b-it-qat-abliterated.Q5_K_M.gguf | 3 ++ gemma-3-4b-it-qat-abliterated.Q5_K_S.gguf | 3 ++ gemma-3-4b-it-qat-abliterated.Q6_K.gguf | 3 ++ gemma-3-4b-it-qat-abliterated.Q8_0.gguf | 3 ++ gemma-3-4b-it-qat-abliterated.f16.gguf | 3 ++ 14 files changed, 147 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 gemma-3-4b-it-qat-abliterated.IQ4_XS.gguf create mode 100644 gemma-3-4b-it-qat-abliterated.Q2_K.gguf create mode 100644 gemma-3-4b-it-qat-abliterated.Q3_K_L.gguf create mode 100644 gemma-3-4b-it-qat-abliterated.Q3_K_M.gguf create mode 100644 gemma-3-4b-it-qat-abliterated.Q3_K_S.gguf create mode 100644 gemma-3-4b-it-qat-abliterated.Q4_K_M.gguf create mode 100644 gemma-3-4b-it-qat-abliterated.Q4_K_S.gguf create mode 100644 gemma-3-4b-it-qat-abliterated.Q5_K_M.gguf create mode 100644 gemma-3-4b-it-qat-abliterated.Q5_K_S.gguf create mode 100644 gemma-3-4b-it-qat-abliterated.Q6_K.gguf create mode 100644 gemma-3-4b-it-qat-abliterated.Q8_0.gguf create mode 100644 gemma-3-4b-it-qat-abliterated.f16.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..76d8f1f --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-4b-it-qat-abliterated.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..a6edf1d --- /dev/null +++ b/README.md @@ -0,0 +1,64 @@ +--- +base_model: mlabonne/gemma-3-4b-it-qat-abliterated +language: +- en +library_name: transformers +license: gemma +quantized_by: mradermacher +--- +## About + + + + + + +static quants of https://huggingface.co/mlabonne/gemma-3-4b-it-qat-abliterated + + +weighted/imatrix quants are available at https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-i1-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.Q2_K.gguf) | Q2_K | 1.8 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.Q3_K_S.gguf) | Q3_K_S | 2.0 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.Q3_K_M.gguf) | Q3_K_M | 2.2 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.Q3_K_L.gguf) | Q3_K_L | 2.3 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.IQ4_XS.gguf) | IQ4_XS | 2.4 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.Q4_K_S.gguf) | Q4_K_S | 2.5 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.Q4_K_M.gguf) | Q4_K_M | 2.6 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.Q5_K_S.gguf) | Q5_K_S | 2.9 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.Q5_K_M.gguf) | Q5_K_M | 2.9 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.Q6_K.gguf) | Q6_K | 3.3 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.Q8_0.gguf) | Q8_0 | 4.2 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-4b-it-qat-abliterated-GGUF/resolve/main/gemma-3-4b-it-qat-abliterated.f16.gguf) | f16 | 7.9 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + + diff --git a/gemma-3-4b-it-qat-abliterated.IQ4_XS.gguf b/gemma-3-4b-it-qat-abliterated.IQ4_XS.gguf new file mode 100644 index 0000000..d211409 --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42967a94a46d2cb12864cf926d94eb80871eb893edc380ad2fb1a650c6616274 +size 2279626368 diff --git a/gemma-3-4b-it-qat-abliterated.Q2_K.gguf b/gemma-3-4b-it-qat-abliterated.Q2_K.gguf new file mode 100644 index 0000000..c2f2584 --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2039043a89940f6c583bfe14362055452015b4abbaf3670440dae039ec48e28e +size 1729164928 diff --git a/gemma-3-4b-it-qat-abliterated.Q3_K_L.gguf b/gemma-3-4b-it-qat-abliterated.Q3_K_L.gguf new file mode 100644 index 0000000..8ff0e17 --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5872cde92a12c81c15fd04bb3be5b30614f025c17c44012ef0545c112857ee5 +size 2236085888 diff --git a/gemma-3-4b-it-qat-abliterated.Q3_K_M.gguf b/gemma-3-4b-it-qat-abliterated.Q3_K_M.gguf new file mode 100644 index 0000000..25bedf7 --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9935fc4e60b33fe77d1b11cc0e9c5c613bb9d7e00eca8bc327ba14e466d90f88 +size 2098460288 diff --git a/gemma-3-4b-it-qat-abliterated.Q3_K_S.gguf b/gemma-3-4b-it-qat-abliterated.Q3_K_S.gguf new file mode 100644 index 0000000..0c25489 --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6de10fe3c99951108ad131e7afaba30410d624620232f5413c1b555a5a764e94 +size 1937364608 diff --git a/gemma-3-4b-it-qat-abliterated.Q4_K_M.gguf b/gemma-3-4b-it-qat-abliterated.Q4_K_M.gguf new file mode 100644 index 0000000..18a92fe --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86e777fd587761daffa74b571de739b4a43406488c309587b52212a7a3f9284d +size 2489894528 diff --git a/gemma-3-4b-it-qat-abliterated.Q4_K_S.gguf b/gemma-3-4b-it-qat-abliterated.Q4_K_S.gguf new file mode 100644 index 0000000..9dfbb03 --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46bed649fd154553aaf4a0a17ac55d3c415942500e7ca07f2e0535ea87951e30 +size 2377930368 diff --git a/gemma-3-4b-it-qat-abliterated.Q5_K_M.gguf b/gemma-3-4b-it-qat-abliterated.Q5_K_M.gguf new file mode 100644 index 0000000..94da0e8 --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8752210eb78806170520d48959fcddebd08324c0ee3d809b42fef320e0109874 +size 2829698688 diff --git a/gemma-3-4b-it-qat-abliterated.Q5_K_S.gguf b/gemma-3-4b-it-qat-abliterated.Q5_K_S.gguf new file mode 100644 index 0000000..29efda8 --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0f494c141a4cdb92125526c1b72fcd5c2180d4da8f1b35891a049e6fa78dd3f +size 2764592768 diff --git a/gemma-3-4b-it-qat-abliterated.Q6_K.gguf b/gemma-3-4b-it-qat-abliterated.Q6_K.gguf new file mode 100644 index 0000000..b4104dd --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29e126c5758eb547b36c9a910582087603ee3b0c3a93f60b04fa11b3dcc8f250 +size 3190740608 diff --git a/gemma-3-4b-it-qat-abliterated.Q8_0.gguf b/gemma-3-4b-it-qat-abliterated.Q8_0.gguf new file mode 100644 index 0000000..5ed01f0 --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4e6b5da5656c4cd11d50952341e017fcf33a5e66a35a29358db662005af58bf +size 4130402688 diff --git a/gemma-3-4b-it-qat-abliterated.f16.gguf b/gemma-3-4b-it-qat-abliterated.f16.gguf new file mode 100644 index 0000000..a514ad6 --- /dev/null +++ b/gemma-3-4b-it-qat-abliterated.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ec4320720b8f32bb9e371a7bea81958db92bcd92493c8596a13706becd47317 +size 7767804288