From 7ed4927a4a3091cbe21ef61adfa1afde8e311f5f Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 21 Apr 2026 13:47:11 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF Source: Original Platform --- .gitattributes | 60 ++++++++++++++ ...3.2-1B-Instruct-abliterated2.i1-IQ1_M.gguf | 3 + ...3.2-1B-Instruct-abliterated2.i1-IQ1_S.gguf | 3 + ...3.2-1B-Instruct-abliterated2.i1-IQ2_M.gguf | 3 + ...3.2-1B-Instruct-abliterated2.i1-IQ2_S.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-IQ2_XS.gguf | 3 + ...2-1B-Instruct-abliterated2.i1-IQ2_XXS.gguf | 3 + ...3.2-1B-Instruct-abliterated2.i1-IQ3_M.gguf | 3 + ...3.2-1B-Instruct-abliterated2.i1-IQ3_S.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-IQ3_XS.gguf | 3 + ...2-1B-Instruct-abliterated2.i1-IQ3_XXS.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-IQ4_NL.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-IQ4_XS.gguf | 3 + ...-3.2-1B-Instruct-abliterated2.i1-Q2_K.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-Q2_K_S.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-Q3_K_L.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-Q3_K_M.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-Q3_K_S.gguf | 3 + ...-3.2-1B-Instruct-abliterated2.i1-Q4_0.gguf | 3 + ...-3.2-1B-Instruct-abliterated2.i1-Q4_1.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-Q4_K_M.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-Q4_K_S.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-Q5_K_M.gguf | 3 + ....2-1B-Instruct-abliterated2.i1-Q5_K_S.gguf | 3 + ...-3.2-1B-Instruct-abliterated2.i1-Q6_K.gguf | 3 + README.md | 81 +++++++++++++++++++ imatrix.dat | 3 + 27 files changed, 216 insertions(+) create mode 100644 .gitattributes create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_M.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_S.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_M.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_S.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XS.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XXS.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_M.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_S.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XS.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XXS.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_NL.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_XS.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K_S.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_L.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_M.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_S.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q4_0.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q4_1.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_M.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_S.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_M.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_S.gguf create mode 100644 Llama-3.2-1B-Instruct-abliterated2.i1-Q6_K.gguf create mode 100644 README.md create mode 100644 imatrix.dat diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..e3bc077 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +imatrix.dat filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-1B-Instruct-abliterated2.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_M.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_M.gguf new file mode 100644 index 0000000..f77ec2f --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7dbdb058093292ff3a33b95625dddd3e127ec9e2efbfe30b626f13f022198f0b +size 413606208 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_S.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_S.gguf new file mode 100644 index 0000000..a04112c --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88c52dee8d0c2e1bc8349ebc46b6bd93250e6ca947896f80b1f78be342799c5c +size 393552192 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_M.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_M.gguf new file mode 100644 index 0000000..9cce45e --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:54c3ed520d260403e6da3689068e78f039f0258401b80759b9f3bcbda54a5fa9 +size 515449152 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_S.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_S.gguf new file mode 100644 index 0000000..a39c436 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d3a767cfbe01287d689ba650be0c7b02204175614626756caf58318538c6518 +size 488710464 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XS.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XS.gguf new file mode 100644 index 0000000..2b4acc2 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c8515c59226c2ab66053cfc932db7aa44bc35c2e2ee987e88da31c4aa5c9d6d +size 475865408 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XXS.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..f506560 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e197d821d3b4768a74bc86e1907c7756ac76ac5882aae40ccc3637bfcd74533 +size 447029568 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_M.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_M.gguf new file mode 100644 index 0000000..cd8e1b8 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8bfd3ab4ce621c2c152d4d147acf962d95465d5c2bc3b36c43ebc10e91ea1945 +size 657289536 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_S.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_S.gguf new file mode 100644 index 0000000..4587c82 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85019a2c47fca38137e8acf052eb3d63364ef5c59944efd2bb8d944eb96fc788 +size 643920192 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XS.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XS.gguf new file mode 100644 index 0000000..7a4caff --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:310cbf290e8624f18369c4d9be2fab2feb7ca307a75612092400b0bcbe0ea5d0 +size 621113664 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XXS.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..f105f11 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d74a004259d952583b3c5622182cf53512fd2aa989433a64124770773cee7a6 +size 562110784 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_NL.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_NL.gguf new file mode 100644 index 0000000..9e3001e --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85fd3b4fbdb78b886d82bb3dc5a4cd5e0a41db5f73471f115d8311c060774498 +size 773026112 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_XS.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_XS.gguf new file mode 100644 index 0000000..7171465 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f03dedc598d5a5f2b6685834d9982adebd2021366d1c9a6bca0ee12f28a00868 +size 743141696 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K.gguf new file mode 100644 index 0000000..712ecb1 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0dee4c0392aba48343240828cdaaf236e9a9ce65e3edce124c3de3df4db10cc +size 580874560 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K_S.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K_S.gguf new file mode 100644 index 0000000..47e0d45 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc27fc351862dd75597817695ae54bd5cce3060426113c7cfdffa8fc336ee3cd +size 554660160 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_L.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_L.gguf new file mode 100644 index 0000000..81b9663 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:276975ea8d1744cf855555d89ed5155c9f62f6ad9c781a257a7b5ac5a37011b4 +size 732524864 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_M.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_M.gguf new file mode 100644 index 0000000..c8d70b3 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4431eb7b02057c54cd597a4e0f80211df4b054cb734be63f34a4d0ee57f069a4 +size 690843968 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_S.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_S.gguf new file mode 100644 index 0000000..7f6eb62 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5632f920ad54de45752fe21cafbe8be2286b1c320f430ec33e23c0609772d86a +size 641691968 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_0.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_0.gguf new file mode 100644 index 0000000..c995ac1 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:562c0fbf886feed6cef9f79395127c90c421f06f29d8ce26de471abd2a86bba1 +size 773026112 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_1.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_1.gguf new file mode 100644 index 0000000..ac238f7 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:098166211578ca6e6040b55a14f539c87e79f31153be8d84d0a9603669919987 +size 831746368 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_M.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_M.gguf new file mode 100644 index 0000000..aefc024 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f3883777cbe6255e82ac6b50278ee6e23471ced12802a8e0b2fc8f6a8d493bc +size 807694656 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_S.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_S.gguf new file mode 100644 index 0000000..b5bbb5c --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:decf3e45444d2f0fc3e322fec3bcc3cb70662563ec07baeecc4d12884241e8cd +size 775647552 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_M.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_M.gguf new file mode 100644 index 0000000..94fab67 --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b6b58fa601cb46ed4723805596ce414f01a655109313b46db6ffdf75f2e902e +size 911503680 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_S.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_S.gguf new file mode 100644 index 0000000..2d6c24e --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2df72fb1b7063cd969107b5e5502067be9bd46ae01c8a441bcc03dc111aac8c7 +size 892563776 diff --git a/Llama-3.2-1B-Instruct-abliterated2.i1-Q6_K.gguf b/Llama-3.2-1B-Instruct-abliterated2.i1-Q6_K.gguf new file mode 100644 index 0000000..c60ddaa --- /dev/null +++ b/Llama-3.2-1B-Instruct-abliterated2.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31b4075053f2e29836527c0dc111848d0feafa7d86f969daa30d8c8f23b500e8 +size 1021800768 diff --git a/README.md b/README.md new file mode 100644 index 0000000..7523fa9 --- /dev/null +++ b/README.md @@ -0,0 +1,81 @@ +--- +base_model: mylesgoose/Llama-3.2-1B-Instruct-abliterated2 +language: +- en +library_name: transformers +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: [] +--- +## About + + + + + + +weighted/imatrix quants of https://huggingface.co/mylesgoose/Llama-3.2-1B-Instruct-abliterated2 + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Llama-3.2-1B-Instruct-abliterated2-i1-GGUF).*** + +static quants are available at https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_S.gguf) | i1-IQ1_S | 0.5 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ1_M.gguf) | i1-IQ1_M | 0.5 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 0.5 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_XS.gguf) | i1-IQ2_XS | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_S.gguf) | i1-IQ2_S | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ2_M.gguf) | i1-IQ2_M | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K_S.gguf) | i1-Q2_K_S | 0.7 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 0.7 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q2_K.gguf) | i1-Q2_K | 0.7 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_XS.gguf) | i1-IQ3_XS | 0.7 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_S.gguf) | i1-Q3_K_S | 0.7 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_S.gguf) | i1-IQ3_S | 0.7 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ3_M.gguf) | i1-IQ3_M | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_M.gguf) | i1-Q3_K_M | 0.8 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q3_K_L.gguf) | i1-Q3_K_L | 0.8 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_XS.gguf) | i1-IQ4_XS | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-IQ4_NL.gguf) | i1-IQ4_NL | 0.9 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_0.gguf) | i1-Q4_0 | 0.9 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_S.gguf) | i1-Q4_K_S | 0.9 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_K_M.gguf) | i1-Q4_K_M | 0.9 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q4_1.gguf) | i1-Q4_1 | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_S.gguf) | i1-Q5_K_S | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q5_K_M.gguf) | i1-Q5_K_M | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.2-1B-Instruct-abliterated2-i1-GGUF/resolve/main/Llama-3.2-1B-Instruct-abliterated2.i1-Q6_K.gguf) | i1-Q6_K | 1.1 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/imatrix.dat b/imatrix.dat new file mode 100644 index 0000000..53171e3 --- /dev/null +++ b/imatrix.dat @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01835e6e6a2cfd29e91425b4da2362b85239a19d51ff13f1bd9a0631510b7d8b +size 1314413