commit 6feee465cc1bc5a8464d322a2565486cdca82e77 Author: ModelHub XC Date: Mon Apr 13 15:37:21 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/gemma-3-1b-it-abliterated-i1-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..023047b --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +imatrix.dat filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-abliterated.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..9970aca --- /dev/null +++ b/README.md @@ -0,0 +1,90 @@ +--- +base_model: huihui-ai/gemma-3-1b-it-abliterated +extra_gated_button_content: Acknowledge license +extra_gated_heading: Access Gemma on Hugging Face +extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and + agree to Google’s usage license. To do this, please ensure you’re logged in to Hugging + Face and click below. Requests are processed immediately. +language: +- en +library_name: transformers +license: gemma +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- chat +- abliterated +- uncensored +--- +## About + + + + + + +weighted/imatrix quants of https://huggingface.co/huihui-ai/gemma-3-1b-it-abliterated + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#gemma-3-1b-it-abliterated-i1-GGUF).*** + +static quants are available at https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ1_S.gguf) | i1-IQ1_S | 0.7 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ1_M.gguf) | i1-IQ1_M | 0.7 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ2_XS.gguf) | i1-IQ2_XS | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ2_S.gguf) | i1-IQ2_S | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ2_M.gguf) | i1-IQ2_M | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q2_K_S.gguf) | i1-Q2_K_S | 0.8 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 0.8 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q3_K_S.gguf) | i1-Q3_K_S | 0.8 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ3_S.gguf) | i1-IQ3_S | 0.8 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ3_XS.gguf) | i1-IQ3_XS | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q2_K.gguf) | i1-Q2_K | 0.8 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ3_M.gguf) | i1-IQ3_M | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ4_XS.gguf) | i1-IQ4_XS | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-IQ4_NL.gguf) | i1-IQ4_NL | 0.8 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q4_0.gguf) | i1-Q4_0 | 0.8 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q3_K_M.gguf) | i1-Q3_K_M | 0.8 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q3_K_L.gguf) | i1-Q3_K_L | 0.9 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q4_1.gguf) | i1-Q4_1 | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q4_K_S.gguf) | i1-Q4_K_S | 0.9 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q4_K_M.gguf) | i1-Q4_K_M | 0.9 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q5_K_S.gguf) | i1-Q5_K_S | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q5_K_M.gguf) | i1-Q5_K_M | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-3-1b-it-abliterated-i1-GGUF/resolve/main/gemma-3-1b-it-abliterated.i1-Q6_K.gguf) | i1-Q6_K | 1.1 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/gemma-3-1b-it-abliterated.i1-IQ1_M.gguf b/gemma-3-1b-it-abliterated.i1-IQ1_M.gguf new file mode 100644 index 0000000..6f9b20e --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72d10667e5b09fd70fd39d94144ca0b2a56187b071239adc59f5e7a3dc0ec925 +size 643486464 diff --git a/gemma-3-1b-it-abliterated.i1-IQ1_S.gguf b/gemma-3-1b-it-abliterated.i1-IQ1_S.gguf new file mode 100644 index 0000000..d955679 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fff86b63399e64f00fd15765c726ab5bbbe4247ef6ceca5f5aeb6c80ed3a7d2 +size 639194112 diff --git a/gemma-3-1b-it-abliterated.i1-IQ2_M.gguf b/gemma-3-1b-it-abliterated.i1-IQ2_M.gguf new file mode 100644 index 0000000..48fe77a --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0ba7276e97f8a8f3577f5e81714ed4c1a1fe98de99d6daafa012426d4c3071d +size 669784320 diff --git a/gemma-3-1b-it-abliterated.i1-IQ2_S.gguf b/gemma-3-1b-it-abliterated.i1-IQ2_S.gguf new file mode 100644 index 0000000..870a429 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:424145aa41d77389a591b0387addf84a4588293345368b61716f1ad5c0058ce6 +size 664061184 diff --git a/gemma-3-1b-it-abliterated.i1-IQ2_XS.gguf b/gemma-3-1b-it-abliterated.i1-IQ2_XS.gguf new file mode 100644 index 0000000..5dbaf19 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81dae21d32f53a21f585f8b24cb6d055e44a0d858d688f2b96695f2128cebdbf +size 657321984 diff --git a/gemma-3-1b-it-abliterated.i1-IQ2_XXS.gguf b/gemma-3-1b-it-abliterated.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..1f3d6c0 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e032988322f470fe6d3d9c2d85fd5551ca1c5fe58ec5df55762716fc66185436 +size 650640384 diff --git a/gemma-3-1b-it-abliterated.i1-IQ3_M.gguf b/gemma-3-1b-it-abliterated.i1-IQ3_M.gguf new file mode 100644 index 0000000..325165c --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:534f2498bfef7737055d88f7b7bf927abdebb53496ab42f4973a152931f731f3 +size 697061376 diff --git a/gemma-3-1b-it-abliterated.i1-IQ3_S.gguf b/gemma-3-1b-it-abliterated.i1-IQ3_S.gguf new file mode 100644 index 0000000..1ab22e4 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76e27957494623ada04b322794fdfeebce004d9dc2c836cc842823fbb7eb216d +size 689815296 diff --git a/gemma-3-1b-it-abliterated.i1-IQ3_XS.gguf b/gemma-3-1b-it-abliterated.i1-IQ3_XS.gguf new file mode 100644 index 0000000..4d2b4cd --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:736b9f0cfcafa28f9754fa191e19780c1d682aee489c80db21163e36101b16ba +size 689815296 diff --git a/gemma-3-1b-it-abliterated.i1-IQ3_XXS.gguf b/gemma-3-1b-it-abliterated.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..05f1a34 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa94862ef1a5ba2587f4cde6301fe790fb8e2b0f16da1fbd62c1b025a07f7118 +size 680110848 diff --git a/gemma-3-1b-it-abliterated.i1-IQ4_NL.gguf b/gemma-3-1b-it-abliterated.i1-IQ4_NL.gguf new file mode 100644 index 0000000..ae46c7f --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b407c2f091b4a3c3917019b6e08d04be857917fec3602660a309cde0a5ccd7c3 +size 721863936 diff --git a/gemma-3-1b-it-abliterated.i1-IQ4_XS.gguf b/gemma-3-1b-it-abliterated.i1-IQ4_XS.gguf new file mode 100644 index 0000000..8959bd5 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:507c6348788bb93d161f5ee238235e14014ed329b45a2cea729e6662851533e9 +size 714435840 diff --git a/gemma-3-1b-it-abliterated.i1-Q2_K.gguf b/gemma-3-1b-it-abliterated.i1-Q2_K.gguf new file mode 100644 index 0000000..6ab9baf --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b81003dd2504c43f5371f78ca73d5c02456e251cc5984ccf7734a3a46d1a9661 +size 689815296 diff --git a/gemma-3-1b-it-abliterated.i1-Q2_K_S.gguf b/gemma-3-1b-it-abliterated.i1-Q2_K_S.gguf new file mode 100644 index 0000000..70b94ab --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de59d341d01d2254e2dc4fb92dfe69b5f7f8715cba03a07746b5547ae0f325f5 +size 671272704 diff --git a/gemma-3-1b-it-abliterated.i1-Q3_K_L.gguf b/gemma-3-1b-it-abliterated.i1-Q3_K_L.gguf new file mode 100644 index 0000000..661b996 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0feac2d3351c176b13cd8d898ab82162a4d00bcdb0933165da6dfa35e2505a29 +size 751576320 diff --git a/gemma-3-1b-it-abliterated.i1-Q3_K_M.gguf b/gemma-3-1b-it-abliterated.i1-Q3_K_M.gguf new file mode 100644 index 0000000..7e6e7db --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47a4f07da6c55b67510e86112eb40fd9e0b5957e4323360279d2915aaf7d2683 +size 722416896 diff --git a/gemma-3-1b-it-abliterated.i1-Q3_K_S.gguf b/gemma-3-1b-it-abliterated.i1-Q3_K_S.gguf new file mode 100644 index 0000000..2c15a6c --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b8e543139e8f675a6889b14331cd2002d4099c320c7c9e970138d50a8438c56 +size 688856832 diff --git a/gemma-3-1b-it-abliterated.i1-Q4_0.gguf b/gemma-3-1b-it-abliterated.i1-Q4_0.gguf new file mode 100644 index 0000000..ce11463 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7704ff7c5b99e9a7d177a604a157e3f32f6eeed5c12d68548d92f4c38613552c +size 721919232 diff --git a/gemma-3-1b-it-abliterated.i1-Q4_1.gguf b/gemma-3-1b-it-abliterated.i1-Q4_1.gguf new file mode 100644 index 0000000..408c360 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc5f4519469c255e36c8ec3b7beac4b78926cf6dccd3b5adbb12a46e8d4f4a5b +size 764036352 diff --git a/gemma-3-1b-it-abliterated.i1-Q4_K_M.gguf b/gemma-3-1b-it-abliterated.i1-Q4_K_M.gguf new file mode 100644 index 0000000..8a580b9 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:188daa1dddee823d9e513793e235c91d26b83ff94dbfa0c7a24ab2eb6b9e72a8 +size 806059008 diff --git a/gemma-3-1b-it-abliterated.i1-Q4_K_S.gguf b/gemma-3-1b-it-abliterated.i1-Q4_K_S.gguf new file mode 100644 index 0000000..1998844 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f6fce5b5c415daf1aaa5bcc21e8616ece0cc78cc2514a25d0f7224803f48bbc +size 780993792 diff --git a/gemma-3-1b-it-abliterated.i1-Q5_K_M.gguf b/gemma-3-1b-it-abliterated.i1-Q5_K_M.gguf new file mode 100644 index 0000000..7f26986 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f84e74af5519ef70438cd1cdd7215da91585eb38fd6df6f05e8cdcc4273a15d6 +size 851346432 diff --git a/gemma-3-1b-it-abliterated.i1-Q5_K_S.gguf b/gemma-3-1b-it-abliterated.i1-Q5_K_S.gguf new file mode 100644 index 0000000..3944e64 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a26b77ab97890dbd0388ebad831ea73ba42d7b14d8bfeca68eca98f998466032 +size 836400384 diff --git a/gemma-3-1b-it-abliterated.i1-Q6_K.gguf b/gemma-3-1b-it-abliterated.i1-Q6_K.gguf new file mode 100644 index 0000000..a1e9890 --- /dev/null +++ b/gemma-3-1b-it-abliterated.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5ee4fc9209e5baed16098a86fb591d452d4c8151046f403eaccc974dae0c6f0 +size 1011739392 diff --git a/imatrix.dat b/imatrix.dat new file mode 100644 index 0000000..e13afa5 --- /dev/null +++ b/imatrix.dat @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ada5282fdcc562250ad068b72c924ebb002198c727c05ade9a352b661e0d349f +size 1430407