From 2b64084bebd5d4bc432b2de712de2735da12f3fe Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sat, 9 May 2026 18:56:18 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/ERank-14B-i1-GGUF Source: Original Platform --- .gitattributes | 60 +++++++++++++++++++++++++++ ERank-14B.i1-IQ1_M.gguf | 3 ++ ERank-14B.i1-IQ1_S.gguf | 3 ++ ERank-14B.i1-IQ2_M.gguf | 3 ++ ERank-14B.i1-IQ2_S.gguf | 3 ++ ERank-14B.i1-IQ2_XS.gguf | 3 ++ ERank-14B.i1-IQ2_XXS.gguf | 3 ++ ERank-14B.i1-IQ3_M.gguf | 3 ++ ERank-14B.i1-IQ3_S.gguf | 3 ++ ERank-14B.i1-IQ3_XS.gguf | 3 ++ ERank-14B.i1-IQ3_XXS.gguf | 3 ++ ERank-14B.i1-IQ4_NL.gguf | 3 ++ ERank-14B.i1-IQ4_XS.gguf | 3 ++ ERank-14B.i1-Q2_K.gguf | 3 ++ ERank-14B.i1-Q2_K_S.gguf | 3 ++ ERank-14B.i1-Q3_K_L.gguf | 3 ++ ERank-14B.i1-Q3_K_M.gguf | 3 ++ ERank-14B.i1-Q3_K_S.gguf | 3 ++ ERank-14B.i1-Q4_0.gguf | 3 ++ ERank-14B.i1-Q4_1.gguf | 3 ++ ERank-14B.i1-Q4_K_M.gguf | 3 ++ ERank-14B.i1-Q4_K_S.gguf | 3 ++ ERank-14B.i1-Q5_K_M.gguf | 3 ++ ERank-14B.i1-Q5_K_S.gguf | 3 ++ ERank-14B.i1-Q6_K.gguf | 3 ++ ERank-14B.imatrix.gguf | 3 ++ README.md | 85 +++++++++++++++++++++++++++++++++++++++ 27 files changed, 220 insertions(+) create mode 100644 .gitattributes create mode 100644 ERank-14B.i1-IQ1_M.gguf create mode 100644 ERank-14B.i1-IQ1_S.gguf create mode 100644 ERank-14B.i1-IQ2_M.gguf create mode 100644 ERank-14B.i1-IQ2_S.gguf create mode 100644 ERank-14B.i1-IQ2_XS.gguf create mode 100644 ERank-14B.i1-IQ2_XXS.gguf create mode 100644 ERank-14B.i1-IQ3_M.gguf create mode 100644 ERank-14B.i1-IQ3_S.gguf create mode 100644 ERank-14B.i1-IQ3_XS.gguf create mode 100644 ERank-14B.i1-IQ3_XXS.gguf create mode 100644 ERank-14B.i1-IQ4_NL.gguf create mode 100644 ERank-14B.i1-IQ4_XS.gguf create mode 100644 ERank-14B.i1-Q2_K.gguf create mode 100644 ERank-14B.i1-Q2_K_S.gguf create mode 100644 ERank-14B.i1-Q3_K_L.gguf create mode 100644 ERank-14B.i1-Q3_K_M.gguf create mode 100644 ERank-14B.i1-Q3_K_S.gguf create mode 100644 ERank-14B.i1-Q4_0.gguf create mode 100644 ERank-14B.i1-Q4_1.gguf create mode 100644 ERank-14B.i1-Q4_K_M.gguf create mode 100644 ERank-14B.i1-Q4_K_S.gguf create mode 100644 ERank-14B.i1-Q5_K_M.gguf create mode 100644 ERank-14B.i1-Q5_K_S.gguf create mode 100644 ERank-14B.i1-Q6_K.gguf create mode 100644 ERank-14B.imatrix.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..695303b --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +ERank-14B.imatrix.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +ERank-14B.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/ERank-14B.i1-IQ1_M.gguf b/ERank-14B.i1-IQ1_M.gguf new file mode 100644 index 0000000..c9fdfe9 --- /dev/null +++ b/ERank-14B.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7fdae6df3692ff952226ebdbf3d8c79e9af0c70447e5482839f1fb9690580e3 +size 3849656224 diff --git a/ERank-14B.i1-IQ1_S.gguf b/ERank-14B.i1-IQ1_S.gguf new file mode 100644 index 0000000..853fdcc --- /dev/null +++ b/ERank-14B.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20d475597b047317d4016bebded9d8b2e393efec6b67a8e98c9ae1edb0c7a30f +size 3579934624 diff --git a/ERank-14B.i1-IQ2_M.gguf b/ERank-14B.i1-IQ2_M.gguf new file mode 100644 index 0000000..01d1aaf --- /dev/null +++ b/ERank-14B.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:199d443d0e4a75113a91998ec602e26a7987ddd123444ae729aecec03bfa7b0a +size 5322941344 diff --git a/ERank-14B.i1-IQ2_S.gguf b/ERank-14B.i1-IQ2_S.gguf new file mode 100644 index 0000000..ece816b --- /dev/null +++ b/ERank-14B.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec1d7063ac917db1aff9f049acbd42bb3e91274af26eee0dab30721c4675f9d2 +size 4963312544 diff --git a/ERank-14B.i1-IQ2_XS.gguf b/ERank-14B.i1-IQ2_XS.gguf new file mode 100644 index 0000000..145578c --- /dev/null +++ b/ERank-14B.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb9241ffdec287fed6de57bee7efdf8bee2a18f62ae81a0b239c545367c7b075 +size 4691589024 diff --git a/ERank-14B.i1-IQ2_XXS.gguf b/ERank-14B.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..b7c73a0 --- /dev/null +++ b/ERank-14B.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21a3858caa4e0891417d58ac71a983720977aba706fc7765c01a7e06b52fa2f0 +size 4299192224 diff --git a/ERank-14B.i1-IQ3_M.gguf b/ERank-14B.i1-IQ3_M.gguf new file mode 100644 index 0000000..e554abe --- /dev/null +++ b/ERank-14B.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f801652eb9917a2703478073b55599a315e2ab7a5b0f2938992ddf925872ceb +size 6883409824 diff --git a/ERank-14B.i1-IQ3_S.gguf b/ERank-14B.i1-IQ3_S.gguf new file mode 100644 index 0000000..d51e8c9 --- /dev/null +++ b/ERank-14B.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6464efcd131ae2248c087f5524d34759cc33b96b73fffddfb34c0767ff2f17a +size 6684958624 diff --git a/ERank-14B.i1-IQ3_XS.gguf b/ERank-14B.i1-IQ3_XS.gguf new file mode 100644 index 0000000..283f020 --- /dev/null +++ b/ERank-14B.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a90f00f9e06c7ffe4e39268ffeb8717d697d92228b5de6f44355ef86f9b3e150 +size 6375301024 diff --git a/ERank-14B.i1-IQ3_XXS.gguf b/ERank-14B.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..bc4147a --- /dev/null +++ b/ERank-14B.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c7e70d5599e5a1be944efe9f161d52d8b8fdac84d5aa210afb6ce6ab91c84f2 +size 5942666144 diff --git a/ERank-14B.i1-IQ4_NL.gguf b/ERank-14B.i1-IQ4_NL.gguf new file mode 100644 index 0000000..5548135 --- /dev/null +++ b/ERank-14B.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d81fe3d167e1c5ae454b8f20d17cc5001a74d24a52fe09977d68de7cb4495a1a +size 8541363104 diff --git a/ERank-14B.i1-IQ4_XS.gguf b/ERank-14B.i1-IQ4_XS.gguf new file mode 100644 index 0000000..db27073 --- /dev/null +++ b/ERank-14B.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3878b7190467ef56c9db5ed541193e012238f7078ab69833f5803a9cf353228c +size 8110730144 diff --git a/ERank-14B.i1-Q2_K.gguf b/ERank-14B.i1-Q2_K.gguf new file mode 100644 index 0000000..426bb61 --- /dev/null +++ b/ERank-14B.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9414c4d51c62f50108eb9306a6efa2ff52a1564743fe9810972584b7bcdc0f62 +size 5753983904 diff --git a/ERank-14B.i1-Q2_K_S.gguf b/ERank-14B.i1-Q2_K_S.gguf new file mode 100644 index 0000000..3caff0f --- /dev/null +++ b/ERank-14B.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d51d45e07cd7ba3dbb7a7b95007b64a4a42de620c1339fc36d9c014360aa6e4d +size 5389849504 diff --git a/ERank-14B.i1-Q3_K_L.gguf b/ERank-14B.i1-Q3_K_L.gguf new file mode 100644 index 0000000..7d95897 --- /dev/null +++ b/ERank-14B.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24579a51be56a1ec446171767deb6bb3d36d009f85df331e41b3d13dfe960c41 +size 7900651424 diff --git a/ERank-14B.i1-Q3_K_M.gguf b/ERank-14B.i1-Q3_K_M.gguf new file mode 100644 index 0000000..497f9f9 --- /dev/null +++ b/ERank-14B.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8994211b9424857e3409d280ccb0b0ec12d8c20a3442a0b7af62f4d46523a546 +size 7321313184 diff --git a/ERank-14B.i1-Q3_K_S.gguf b/ERank-14B.i1-Q3_K_S.gguf new file mode 100644 index 0000000..b72e258 --- /dev/null +++ b/ERank-14B.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88f2a3a1d2b71359613beb957dfd3a4af06f25e44c397cf24138e412fdda3126 +size 6657105824 diff --git a/ERank-14B.i1-Q4_0.gguf b/ERank-14B.i1-Q4_0.gguf new file mode 100644 index 0000000..c423c68 --- /dev/null +++ b/ERank-14B.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a073332a765d5b2873a4d073d461775206a73c540e3a0b54a9c92c7dd1a9add +size 8543001504 diff --git a/ERank-14B.i1-Q4_1.gguf b/ERank-14B.i1-Q4_1.gguf new file mode 100644 index 0000000..4bc3caf --- /dev/null +++ b/ERank-14B.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5cc535e246619c517f735b09e24c20463b45415f5c0404060770fc0542997d0 +size 9389521824 diff --git a/ERank-14B.i1-Q4_K_M.gguf b/ERank-14B.i1-Q4_K_M.gguf new file mode 100644 index 0000000..18cb713 --- /dev/null +++ b/ERank-14B.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:307b1e5f59b17a317ae6d212fa874f3178bf4aa81b18319966ac71d21350c92d +size 9001753504 diff --git a/ERank-14B.i1-Q4_K_S.gguf b/ERank-14B.i1-Q4_K_S.gguf new file mode 100644 index 0000000..859eba3 --- /dev/null +++ b/ERank-14B.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c769a796a5bcf370ec670ddbb991976fd1623a7375724eadc25688426993bcd6 +size 8573475744 diff --git a/ERank-14B.i1-Q5_K_M.gguf b/ERank-14B.i1-Q5_K_M.gguf new file mode 100644 index 0000000..5a8cba5 --- /dev/null +++ b/ERank-14B.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a4ca0d0b3bd10d3e0aec5349cfcadf925d050668cf35aed4c3e3f2f8f59dd26 +size 10514570144 diff --git a/ERank-14B.i1-Q5_K_S.gguf b/ERank-14B.i1-Q5_K_S.gguf new file mode 100644 index 0000000..4ff3259 --- /dev/null +++ b/ERank-14B.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73f5ef25ee1b62a5a73a822cbc9d7435f7970398db5b989595c46ced5bc315c9 +size 10263894944 diff --git a/ERank-14B.i1-Q6_K.gguf b/ERank-14B.i1-Q6_K.gguf new file mode 100644 index 0000000..7230ee7 --- /dev/null +++ b/ERank-14B.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4db324d140c2b0ab8e5751379645a918d66bbfe4300f219e319168ca26ffd432 +size 12121937824 diff --git a/ERank-14B.imatrix.gguf b/ERank-14B.imatrix.gguf new file mode 100644 index 0000000..39e3fc3 --- /dev/null +++ b/ERank-14B.imatrix.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aba328a9db0d0ec2d7b300504e843a5c5c70cc49b31b6591be721dfaad500083 +size 7743552 diff --git a/README.md b/README.md new file mode 100644 index 0000000..7801add --- /dev/null +++ b/README.md @@ -0,0 +1,85 @@ +--- +base_model: Alibaba-NLP/ERank-14B +language: +- en +library_name: transformers +license: apache-2.0 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +--- +## About + + + + + + + + + +weighted/imatrix quants of https://huggingface.co/Alibaba-NLP/ERank-14B + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#ERank-14B-i1-GGUF).*** + +static quants are available at https://huggingface.co/mradermacher/ERank-14B-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.imatrix.gguf) | imatrix | 0.1 | imatrix file (for creating your own quants) | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ1_S.gguf) | i1-IQ1_S | 3.7 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ1_M.gguf) | i1-IQ1_M | 3.9 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 4.4 | | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ2_XS.gguf) | i1-IQ2_XS | 4.8 | | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ2_S.gguf) | i1-IQ2_S | 5.1 | | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ2_M.gguf) | i1-IQ2_M | 5.4 | | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q2_K_S.gguf) | i1-Q2_K_S | 5.5 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q2_K.gguf) | i1-Q2_K | 5.9 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 6.0 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ3_XS.gguf) | i1-IQ3_XS | 6.5 | | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q3_K_S.gguf) | i1-Q3_K_S | 6.8 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ3_S.gguf) | i1-IQ3_S | 6.8 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ3_M.gguf) | i1-IQ3_M | 7.0 | | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q3_K_M.gguf) | i1-Q3_K_M | 7.4 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q3_K_L.gguf) | i1-Q3_K_L | 8.0 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ4_XS.gguf) | i1-IQ4_XS | 8.2 | | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-IQ4_NL.gguf) | i1-IQ4_NL | 8.6 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q4_0.gguf) | i1-Q4_0 | 8.6 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q4_K_S.gguf) | i1-Q4_K_S | 8.7 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q4_K_M.gguf) | i1-Q4_K_M | 9.1 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q4_1.gguf) | i1-Q4_1 | 9.5 | | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q5_K_S.gguf) | i1-Q5_K_S | 10.4 | | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q5_K_M.gguf) | i1-Q5_K_M | 10.6 | | +| [GGUF](https://huggingface.co/mradermacher/ERank-14B-i1-GGUF/resolve/main/ERank-14B.i1-Q6_K.gguf) | i1-Q6_K | 12.2 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + +