From 95b402e76a1c267630fe62f4b853efc28a5bd527 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sun, 3 May 2026 10:04:40 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF Source: Original Platform --- .gitattributes | 60 +++++++++++++ README.md | 88 +++++++++++++++++++ ...a-2-2b-jpn-it-abliterated-18.i1-IQ1_M.gguf | 3 + ...a-2-2b-jpn-it-abliterated-18.i1-IQ1_S.gguf | 3 + ...a-2-2b-jpn-it-abliterated-18.i1-IQ2_M.gguf | 3 + ...a-2-2b-jpn-it-abliterated-18.i1-IQ2_S.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-IQ2_XS.gguf | 3 + ...2-2b-jpn-it-abliterated-18.i1-IQ2_XXS.gguf | 3 + ...a-2-2b-jpn-it-abliterated-18.i1-IQ3_M.gguf | 3 + ...a-2-2b-jpn-it-abliterated-18.i1-IQ3_S.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-IQ3_XS.gguf | 3 + ...2-2b-jpn-it-abliterated-18.i1-IQ3_XXS.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-IQ4_NL.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-IQ4_XS.gguf | 3 + gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-Q2_K_S.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-Q3_K_L.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-Q3_K_M.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-Q3_K_S.gguf | 3 + gemma-2-2b-jpn-it-abliterated-18.i1-Q4_0.gguf | 3 + gemma-2-2b-jpn-it-abliterated-18.i1-Q4_1.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-Q4_K_M.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-Q4_K_S.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-Q5_K_M.gguf | 3 + ...-2-2b-jpn-it-abliterated-18.i1-Q5_K_S.gguf | 3 + gemma-2-2b-jpn-it-abliterated-18.i1-Q6_K.gguf | 3 + imatrix.dat | 3 + 27 files changed, 223 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_M.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_S.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_M.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_S.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XS.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XXS.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_M.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_S.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XS.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XXS.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_NL.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_XS.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K_S.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_L.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_M.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_S.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q4_0.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q4_1.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_M.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_S.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_M.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_S.gguf create mode 100644 gemma-2-2b-jpn-it-abliterated-18.i1-Q6_K.gguf create mode 100644 imatrix.dat diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..c818e7e --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +imatrix.dat filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +gemma-2-2b-jpn-it-abliterated-18.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..4c6ffe9 --- /dev/null +++ b/README.md @@ -0,0 +1,88 @@ +--- +base_model: ymcki/gemma-2-2b-jpn-it-abliterated-18 +datasets: +- mlabonne/harmless_alpaca +- mlabonne/harmful_behaviors +language: +- multilingual +library_name: transformers +license: gemma +license_link: https://ai.google.dev/gemma/terms +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- nlp +- code +--- +## About + + + + + + +weighted/imatrix quants of https://huggingface.co/ymcki/gemma-2-2b-jpn-it-abliterated-18 + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#gemma-2-2b-jpn-it-abliterated-18-i1-GGUF).*** + +static quants are available at https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_S.gguf) | i1-IQ1_S | 0.9 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_M.gguf) | i1-IQ1_M | 1.0 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XS.gguf) | i1-IQ2_XS | 1.1 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_S.gguf) | i1-IQ2_S | 1.1 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_M.gguf) | i1-IQ2_M | 1.2 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K_S.gguf) | i1-Q2_K_S | 1.3 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 1.3 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K.gguf) | i1-Q2_K | 1.3 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XS.gguf) | i1-IQ3_XS | 1.4 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_S.gguf) | i1-IQ3_S | 1.5 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_S.gguf) | i1-Q3_K_S | 1.5 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_M.gguf) | i1-IQ3_M | 1.5 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_M.gguf) | i1-Q3_K_M | 1.6 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_L.gguf) | i1-Q3_K_L | 1.7 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_XS.gguf) | i1-IQ4_XS | 1.7 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_NL.gguf) | i1-IQ4_NL | 1.7 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_0.gguf) | i1-Q4_0 | 1.7 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_S.gguf) | i1-Q4_K_S | 1.7 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_M.gguf) | i1-Q4_K_M | 1.8 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_1.gguf) | i1-Q4_1 | 1.9 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_S.gguf) | i1-Q5_K_S | 2.0 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_M.gguf) | i1-Q5_K_M | 2.0 | | +| [GGUF](https://huggingface.co/mradermacher/gemma-2-2b-jpn-it-abliterated-18-i1-GGUF/resolve/main/gemma-2-2b-jpn-it-abliterated-18.i1-Q6_K.gguf) | i1-Q6_K | 2.3 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_M.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_M.gguf new file mode 100644 index 0000000..26ee623 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7ce06d764c1c5f8c89b89df3a6acfee105875bd868d8850fedf029b2d473111 +size 873798880 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_S.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_S.gguf new file mode 100644 index 0000000..f085d63 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc23cca67a492d96ae4abd428cf3f2d6cd838dbc3868de389fad90075df3bf8a +size 832160992 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_M.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_M.gguf new file mode 100644 index 0000000..e78c4bc --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bfcb1bcbeedcedd217a3e39c21dbe09c7e605ef8a20fd7e7cfa14e32223fec57 +size 1088015584 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_S.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_S.gguf new file mode 100644 index 0000000..764d918 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc0027a38ffc24d7c2e5078017e82f2df94cda6929edb7a5f182b6dcd8b3d4d6 +size 1032498400 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XS.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XS.gguf new file mode 100644 index 0000000..df3e6d1 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f104f3ff74591e543ecd3c0db2a33ea7ec8725dfc9bd636d892c5c84d87c878 +size 1002546400 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XXS.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..65d86e6 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e4851aba23ffb99f71b8d418195471e69a9bfc78827fa08802105ced6d2f09e +size 943195360 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_M.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_M.gguf new file mode 100644 index 0000000..df8bac5 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c1eebbcbd977636718135b6301a176e60cd5a8fce4af2e0d82a8aa78c6e7970 +size 1393562848 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_S.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_S.gguf new file mode 100644 index 0000000..29bb027 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb12a8f3dee6d0ed0167dc3b6c97de0edc027341ded05e60dd94b32daa58dc89 +size 1360661728 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XS.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XS.gguf new file mode 100644 index 0000000..2665156 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f515d13c32969d3c14316d3ed270360beb5b3df01c802837b75e8ac109d8a56e +size 1314213088 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XXS.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..fa49d8c --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aadf05167b7a295f9519a32e0b901941db160467d61537abd0ed1d2d63a3b2ae +size 1181687008 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_NL.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_NL.gguf new file mode 100644 index 0000000..041af45 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:419194d2f3891c6c78f6dd854b0cbcbdfca125632e6aef0dec31537c8ee34475 +size 1629510880 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_XS.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_XS.gguf new file mode 100644 index 0000000..e125757 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52df7204f4950ccc66fcf3c53f24a27d162ea3d4b94eb8a0aacb16b7909d1108 +size 1566252256 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K.gguf new file mode 100644 index 0000000..59538bc --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b133bfe794ebb50e3d198e9fea66e46542928ac1ed193bc0d7543c0efd3764d0 +size 1229831392 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K_S.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K_S.gguf new file mode 100644 index 0000000..bf6ad1a --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc71c0dabc9b0ce715469632ed2390db9ab7244ed4d4b2b2754e64ff3fd4357b +size 1170001120 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_L.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_L.gguf new file mode 100644 index 0000000..26dc411 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:923834316ad253aeed23f60ce4b3549994c2000f5ab9802e28b27859d489e83c +size 1550437600 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_M.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_M.gguf new file mode 100644 index 0000000..f03798e --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee1a8b6907fd443ab7dd744936ac4b006126ac7b7d29c5de52c70196ee641873 +size 1461669088 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_S.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_S.gguf new file mode 100644 index 0000000..c3b229d --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23b7cc60aed12b9e8be375df17f5c1c616c3f5f0a250e5c0f72acd72a8814da3 +size 1360661728 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_0.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_0.gguf new file mode 100644 index 0000000..8c11b07 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f933fd29fb1cf9b822f087036dd286fbbfd036691fb106759063fe43a81d0880 +size 1633492192 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_1.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_1.gguf new file mode 100644 index 0000000..7357e5a --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1758d5dc33b5086a1bd95dca5a7d3fba83c3a5a5fe8369b6bb9ecfc34c666f62 +size 1756028128 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_M.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_M.gguf new file mode 100644 index 0000000..33cca6e --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e40fee5cb08d26b5f4d6be2af534fa051a5bce27ac1c6c3b47db9335b482575 +size 1708584160 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_S.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_S.gguf new file mode 100644 index 0000000..cc7051c --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:240dcf9a356385f314cbc953a7fbaaf1422cb6b23d5b517dfbef993b62efc14e +size 1638653152 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_M.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_M.gguf new file mode 100644 index 0000000..3c2d6da --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae91d947ffa3510efc999ccaa099221a75a1c4d9ff2db087516f86a49d6ae0e1 +size 1923280096 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_S.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_S.gguf new file mode 100644 index 0000000..f2094ec --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0cb6dfb09cf9eb519930602c7eb217c201f6431260820dc8a53ea1cefa655510 +size 1882545376 diff --git a/gemma-2-2b-jpn-it-abliterated-18.i1-Q6_K.gguf b/gemma-2-2b-jpn-it-abliterated-18.i1-Q6_K.gguf new file mode 100644 index 0000000..6278190 --- /dev/null +++ b/gemma-2-2b-jpn-it-abliterated-18.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:446c5a77be18e4fcce9ab0e8946b5fb436db09c6608c6dc63a1f32aa46ee6c0c +size 2151394528 diff --git a/imatrix.dat b/imatrix.dat new file mode 100644 index 0000000..fc10969 --- /dev/null +++ b/imatrix.dat @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f609492ad67852ae9a7887208b68faddf6368348a2e98316d09955513de4a881 +size 2375559