From 99520d1d9aca2b3f2328df81fa69d4bcec870cc1 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Fri, 8 May 2026 15:18:58 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF Source: Original Platform --- .gitattributes | 58 +++++++++++++++++++++ 14B-Qwen2.5-Kunou-v1.i1-IQ1_M.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-IQ1_S.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-IQ2_M.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-IQ2_S.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-IQ2_XS.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-IQ2_XXS.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-IQ3_M.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-IQ3_S.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-IQ3_XS.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-IQ3_XXS.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-IQ4_XS.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-Q2_K.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-Q2_K_S.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-Q3_K_L.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-Q3_K_M.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-Q3_K_S.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-Q4_0.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-Q4_K_M.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-Q4_K_S.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-Q5_K_M.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-Q5_K_S.gguf | 3 ++ 14B-Qwen2.5-Kunou-v1.i1-Q6_K.gguf | 3 ++ README.md | 78 ++++++++++++++++++++++++++++ imatrix.dat | 3 ++ 25 files changed, 205 insertions(+) create mode 100644 .gitattributes create mode 100644 14B-Qwen2.5-Kunou-v1.i1-IQ1_M.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-IQ1_S.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-IQ2_M.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-IQ2_S.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-IQ2_XS.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-IQ2_XXS.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-IQ3_M.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-IQ3_S.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-IQ3_XS.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-IQ3_XXS.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-IQ4_XS.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-Q2_K.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-Q2_K_S.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-Q3_K_L.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-Q3_K_M.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-Q3_K_S.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-Q4_0.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-Q4_K_M.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-Q4_K_S.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-Q5_K_M.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-Q5_K_S.gguf create mode 100644 14B-Qwen2.5-Kunou-v1.i1-Q6_K.gguf create mode 100644 README.md create mode 100644 imatrix.dat diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..61a4fe1 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,58 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +imatrix.dat filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +14B-Qwen2.5-Kunou-v1.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/14B-Qwen2.5-Kunou-v1.i1-IQ1_M.gguf b/14B-Qwen2.5-Kunou-v1.i1-IQ1_M.gguf new file mode 100644 index 0000000..705be73 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bed8396dc30a9fa1f6b7b6b0a801825d4dcd8f14cf8ff2c746f7374cf19ce975 +size 3872309792 diff --git a/14B-Qwen2.5-Kunou-v1.i1-IQ1_S.gguf b/14B-Qwen2.5-Kunou-v1.i1-IQ1_S.gguf new file mode 100644 index 0000000..83e26b8 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47f042ab1fe5732ef982327c44661438421a53f3f3cf874ca7605d8d4e959e9b +size 3607994912 diff --git a/14B-Qwen2.5-Kunou-v1.i1-IQ2_M.gguf b/14B-Qwen2.5-Kunou-v1.i1-IQ2_M.gguf new file mode 100644 index 0000000..0265443 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84dde7f6a95ed0dc25427b6d41686ec7c8d78b722ea6ce2e2a9f3dcd7f5f7e7a +size 5356147232 diff --git a/14B-Qwen2.5-Kunou-v1.i1-IQ2_S.gguf b/14B-Qwen2.5-Kunou-v1.i1-IQ2_S.gguf new file mode 100644 index 0000000..55b8642 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95cb573e0eb52759ee2c8fa5d334c57501e6c15752fb16ba146f5adac7db0a21 +size 5003727392 diff --git a/14B-Qwen2.5-Kunou-v1.i1-IQ2_XS.gguf b/14B-Qwen2.5-Kunou-v1.i1-IQ2_XS.gguf new file mode 100644 index 0000000..74591fc --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d383212d686f956dd64ccc59f4ff722520ffac36face4d0aa7f39effbeecd34 +size 4704576032 diff --git a/14B-Qwen2.5-Kunou-v1.i1-IQ2_XXS.gguf b/14B-Qwen2.5-Kunou-v1.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..4f9caea --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e0f871a03fad0f88155684bade69f7a87fe55310597ee405bd2e65b4324c889 +size 4312834592 diff --git a/14B-Qwen2.5-Kunou-v1.i1-IQ3_M.gguf b/14B-Qwen2.5-Kunou-v1.i1-IQ3_M.gguf new file mode 100644 index 0000000..eb4c9a0 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa52545d80aefb3b6873fb3bd2313d050e453d5b3601656e2cd20fd8b80b5298 +size 6916538912 diff --git a/14B-Qwen2.5-Kunou-v1.i1-IQ3_S.gguf b/14B-Qwen2.5-Kunou-v1.i1-IQ3_S.gguf new file mode 100644 index 0000000..075ec77 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4b5cca91b76636bc7453176c2e43f5f2c1dc8d641bbe8b783d88ece866b117f +size 6693020192 diff --git a/14B-Qwen2.5-Kunou-v1.i1-IQ3_XS.gguf b/14B-Qwen2.5-Kunou-v1.i1-IQ3_XS.gguf new file mode 100644 index 0000000..cd00e05 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40d78abbc372b2cdcecfbfe4bebccf6934ada23175c5afdb40ee4b26329ac41b +size 6383362592 diff --git a/14B-Qwen2.5-Kunou-v1.i1-IQ3_XXS.gguf b/14B-Qwen2.5-Kunou-v1.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..1a6f3e9 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68e698e907e5a02573c647cbeb2b6d5ffd9c698775e07c226fe6ed516dfbf6f9 +size 5946708512 diff --git a/14B-Qwen2.5-Kunou-v1.i1-IQ4_XS.gguf b/14B-Qwen2.5-Kunou-v1.i1-IQ4_XS.gguf new file mode 100644 index 0000000..a49943e --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d969bceed48b5413ca2f33ec3d7292a2cab94884e46f0b6f1b6315a737527b67 +size 8119841312 diff --git a/14B-Qwen2.5-Kunou-v1.i1-Q2_K.gguf b/14B-Qwen2.5-Kunou-v1.i1-Q2_K.gguf new file mode 100644 index 0000000..bb4c6d1 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2624882c939e4a7e29f0a9d5f746ce93489b7694f8185de2af17866f21989d8e +size 5770498592 diff --git a/14B-Qwen2.5-Kunou-v1.i1-Q2_K_S.gguf b/14B-Qwen2.5-Kunou-v1.i1-Q2_K_S.gguf new file mode 100644 index 0000000..c4bb3e5 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7104eb7bac6dae0b33c7e8ede1ba74622ff93147b2174b88948a48c0192b3cce +size 5397189152 diff --git a/14B-Qwen2.5-Kunou-v1.i1-Q3_K_L.gguf b/14B-Qwen2.5-Kunou-v1.i1-Q3_K_L.gguf new file mode 100644 index 0000000..b2723f4 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84de94fad81cce683b986cfc75a279921602f793dcdfc94795fce5d1855f6371 +size 7924769312 diff --git a/14B-Qwen2.5-Kunou-v1.i1-Q3_K_M.gguf b/14B-Qwen2.5-Kunou-v1.i1-Q3_K_M.gguf new file mode 100644 index 0000000..76e269f --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e6bad85e6eab36dabf301ce4c2739675c56a28a4a0be95b13f85377c20c455f +size 7339205152 diff --git a/14B-Qwen2.5-Kunou-v1.i1-Q3_K_S.gguf b/14B-Qwen2.5-Kunou-v1.i1-Q3_K_S.gguf new file mode 100644 index 0000000..1512cbe --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24b140baa6039e22a40a05263a7014923d4381d3dfd23846be5ce4b4a09ea043 +size 6659596832 diff --git a/14B-Qwen2.5-Kunou-v1.i1-Q4_0.gguf b/14B-Qwen2.5-Kunou-v1.i1-Q4_0.gguf new file mode 100644 index 0000000..efe520d --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a34c61ca8faf30e4366ba50adbbbee5f43d0898f9d8bf24a6de6b7f7b55945c +size 8544268832 diff --git a/14B-Qwen2.5-Kunou-v1.i1-Q4_K_M.gguf b/14B-Qwen2.5-Kunou-v1.i1-Q4_K_M.gguf new file mode 100644 index 0000000..b1ca279 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45a506cddb40719a5fef8a7fb630db1ef7a268cbe68098bc857003dde29d016d +size 8988111392 diff --git a/14B-Qwen2.5-Kunou-v1.i1-Q4_K_S.gguf b/14B-Qwen2.5-Kunou-v1.i1-Q4_K_S.gguf new file mode 100644 index 0000000..6c1457c --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d35c5bbac8d20961fd718b164662c9f3c62e34e345db2aa3037f0a62dcfbc57 +size 8573432352 diff --git a/14B-Qwen2.5-Kunou-v1.i1-Q5_K_M.gguf b/14B-Qwen2.5-Kunou-v1.i1-Q5_K_M.gguf new file mode 100644 index 0000000..84c5a65 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9eec73b59997e6deca296aabd2c56f106850793a0ef68e7ff245f11d828986e8 +size 10508874272 diff --git a/14B-Qwen2.5-Kunou-v1.i1-Q5_K_S.gguf b/14B-Qwen2.5-Kunou-v1.i1-Q5_K_S.gguf new file mode 100644 index 0000000..882c328 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74dce3df032035adc3147ad1a9f5aeca44da54c41685237d1100208e47fb241c +size 10266554912 diff --git a/14B-Qwen2.5-Kunou-v1.i1-Q6_K.gguf b/14B-Qwen2.5-Kunou-v1.i1-Q6_K.gguf new file mode 100644 index 0000000..dc73eb2 --- /dev/null +++ b/14B-Qwen2.5-Kunou-v1.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67bb0c8022d5fb8f5e102c8b3d278787f2605a4737e8eec0704938a5ca6722f2 +size 12124684832 diff --git a/README.md b/README.md new file mode 100644 index 0000000..892bd6f --- /dev/null +++ b/README.md @@ -0,0 +1,78 @@ +--- +base_model: Sao10K/14B-Qwen2.5-Kunou-v1 +language: +- en +library_name: transformers +license: other +license_link: https://huggingface.co/Qwen/Qwen2.5-14B-Instruct/blob/main/LICENSE +license_name: qwen +quantized_by: mradermacher +tags: +- generated_from_trainer +--- +## About + + + + + + +weighted/imatrix quants of https://huggingface.co/Sao10K/14B-Qwen2.5-Kunou-v1 + + +static quants are available at https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-IQ1_S.gguf) | i1-IQ1_S | 3.7 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-IQ1_M.gguf) | i1-IQ1_M | 4.0 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 4.4 | | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-IQ2_XS.gguf) | i1-IQ2_XS | 4.8 | | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-IQ2_S.gguf) | i1-IQ2_S | 5.1 | | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-IQ2_M.gguf) | i1-IQ2_M | 5.5 | | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-Q2_K_S.gguf) | i1-Q2_K_S | 5.5 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-Q2_K.gguf) | i1-Q2_K | 5.9 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 6.0 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-IQ3_XS.gguf) | i1-IQ3_XS | 6.5 | | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-Q3_K_S.gguf) | i1-Q3_K_S | 6.8 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-IQ3_S.gguf) | i1-IQ3_S | 6.8 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-IQ3_M.gguf) | i1-IQ3_M | 7.0 | | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-Q3_K_M.gguf) | i1-Q3_K_M | 7.4 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-Q3_K_L.gguf) | i1-Q3_K_L | 8.0 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-IQ4_XS.gguf) | i1-IQ4_XS | 8.2 | | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-Q4_0.gguf) | i1-Q4_0 | 8.6 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-Q4_K_S.gguf) | i1-Q4_K_S | 8.7 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-Q4_K_M.gguf) | i1-Q4_K_M | 9.1 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-Q5_K_S.gguf) | i1-Q5_K_S | 10.4 | | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-Q5_K_M.gguf) | i1-Q5_K_M | 10.6 | | +| [GGUF](https://huggingface.co/mradermacher/14B-Qwen2.5-Kunou-v1-i1-GGUF/resolve/main/14B-Qwen2.5-Kunou-v1.i1-Q6_K.gguf) | i1-Q6_K | 12.2 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/imatrix.dat b/imatrix.dat new file mode 100644 index 0000000..5a6ac06 --- /dev/null +++ b/imatrix.dat @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3316bc25d718a74f5e148be1cb361d93d184c97bf92111db6e330bd242a27970 +size 8563597