commit d8c7a52db526f1486b06e48a3d71144e9f31cf26 Author: ModelHub XC Date: Fri Apr 10 19:20:54 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/Chekhov-24B-v1.0-i1-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..b171db7 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,59 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.imatrix.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +Chekhov-24B-v1.0.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Chekhov-24B-v1.0.i1-IQ1_M.gguf b/Chekhov-24B-v1.0.i1-IQ1_M.gguf new file mode 100644 index 0000000..fdcd694 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:943eecf5f6b5a70320ea1fe355b4b043439c6a66c1925f122a6e4262cf16ce84 +size 5750497632 diff --git a/Chekhov-24B-v1.0.i1-IQ1_S.gguf b/Chekhov-24B-v1.0.i1-IQ1_S.gguf new file mode 100644 index 0000000..3f3b80c --- /dev/null +++ b/Chekhov-24B-v1.0.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd9127fa83c65c55c49c78de2a5f3099f9460ea58b6361081df9e1eed72b76b4 +size 5273723232 diff --git a/Chekhov-24B-v1.0.i1-IQ2_M.gguf b/Chekhov-24B-v1.0.i1-IQ2_M.gguf new file mode 100644 index 0000000..e4e0bb3 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04dcc9457eaae436fbbb6f74ade8e45c1305ed0e672d91f0efd2371d0e372648 +size 8114053472 diff --git a/Chekhov-24B-v1.0.i1-IQ2_S.gguf b/Chekhov-24B-v1.0.i1-IQ2_S.gguf new file mode 100644 index 0000000..67a0c62 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d09e7fa7d603e1d0f053812210723f4f4aeea596b4eb66fa9df0b11b8646c40a +size 7478354272 diff --git a/Chekhov-24B-v1.0.i1-IQ2_XS.gguf b/Chekhov-24B-v1.0.i1-IQ2_XS.gguf new file mode 100644 index 0000000..b39e1a4 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6604a5c8f3f646eef4641480b41eb43258bb0d4e4e2f0d96174b916427173ac +size 7207035232 diff --git a/Chekhov-24B-v1.0.i1-IQ2_XXS.gguf b/Chekhov-24B-v1.0.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..b9c65ab --- /dev/null +++ b/Chekhov-24B-v1.0.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:990b5d218733a4344abc8816f08bff2a097ad7a90f7841647fd778548911d71c +size 6545121632 diff --git a/Chekhov-24B-v1.0.i1-IQ3_M.gguf b/Chekhov-24B-v1.0.i1-IQ3_M.gguf new file mode 100644 index 0000000..cc2bea4 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d5db1d1b5b9de7365baafc7fb1c1ca866f411d8963d4ca8bee6a6c5fd47242e2 +size 10650952032 diff --git a/Chekhov-24B-v1.0.i1-IQ3_S.gguf b/Chekhov-24B-v1.0.i1-IQ3_S.gguf new file mode 100644 index 0000000..1cb5960 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4a96b08d15b8426ed5b5074443e4daf80a5ee6c9fd11fd90b690f76bd511e2b +size 10428129632 diff --git a/Chekhov-24B-v1.0.i1-IQ3_XS.gguf b/Chekhov-24B-v1.0.i1-IQ3_XS.gguf new file mode 100644 index 0000000..18ede7e --- /dev/null +++ b/Chekhov-24B-v1.0.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6dcf0b91e028943463c38d49749aebd3ecc7f7415ac94be8c8ec24fb5d2d12f1 +size 9907118432 diff --git a/Chekhov-24B-v1.0.i1-IQ3_XXS.gguf b/Chekhov-24B-v1.0.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..5d053de --- /dev/null +++ b/Chekhov-24B-v1.0.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3414276711d143434f196462c0e4065363d5fddbdd0f57010aee378e03a166e +size 9280594272 diff --git a/Chekhov-24B-v1.0.i1-IQ4_XS.gguf b/Chekhov-24B-v1.0.i1-IQ4_XS.gguf new file mode 100644 index 0000000..81aa8a0 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe83861fb22b6a4ef1c606820a3523646fcfcc244a9215fb9b3c2d71bb84b3b1 +size 12758917472 diff --git a/Chekhov-24B-v1.0.i1-Q2_K.gguf b/Chekhov-24B-v1.0.i1-Q2_K.gguf new file mode 100644 index 0000000..a7f92db --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13f8bff54fbeab557ecf3e5f9e8413103368f1d384d39d845704c16c10bb7c54 +size 8890327392 diff --git a/Chekhov-24B-v1.0.i1-Q2_K_S.gguf b/Chekhov-24B-v1.0.i1-Q2_K_S.gguf new file mode 100644 index 0000000..22a488e --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:826c81221bd2a4a5c9ea8644f94bb633362c5eb6667c0e06257e4ee340718dca +size 8320164192 diff --git a/Chekhov-24B-v1.0.i1-Q3_K_L.gguf b/Chekhov-24B-v1.0.i1-Q3_K_L.gguf new file mode 100644 index 0000000..5342833 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b8034eaa03c6fe5b56de6b065f3899b8c470109dd0edc46202f553fe2aa301e +size 12400763232 diff --git a/Chekhov-24B-v1.0.i1-Q3_K_M.gguf b/Chekhov-24B-v1.0.i1-Q3_K_M.gguf new file mode 100644 index 0000000..791ab84 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65e8b0a9e476b905eb58adc9ed308101a2ad91773bd947fe29c0790309f3a5a5 +size 11474084192 diff --git a/Chekhov-24B-v1.0.i1-Q3_K_S.gguf b/Chekhov-24B-v1.0.i1-Q3_K_S.gguf new file mode 100644 index 0000000..108257c --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afcb1187f870163a7304642db76003663a37fa099453286c562e9482e31423b6 +size 10400276832 diff --git a/Chekhov-24B-v1.0.i1-Q4_0.gguf b/Chekhov-24B-v1.0.i1-Q4_0.gguf new file mode 100644 index 0000000..8d5a56b --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb00a1bc409a7433cb58b851da405452c531d08d2fe7afb3a91388d6532e70c6 +size 13494231392 diff --git a/Chekhov-24B-v1.0.i1-Q4_1.gguf b/Chekhov-24B-v1.0.i1-Q4_1.gguf new file mode 100644 index 0000000..efa06f5 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40cef8429a0f8339a9a63e36e533e2d6bbf777141e2a03cf29f5054226a0ed89 +size 14873108832 diff --git a/Chekhov-24B-v1.0.i1-Q4_K_M.gguf b/Chekhov-24B-v1.0.i1-Q4_K_M.gguf new file mode 100644 index 0000000..1b1c4f1 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c47c8dfa071d058869182a1e305eca5382e3cbbab727052eced05aa91399173 +size 14333911392 diff --git a/Chekhov-24B-v1.0.i1-Q4_K_S.gguf b/Chekhov-24B-v1.0.i1-Q4_K_S.gguf new file mode 100644 index 0000000..125496a --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5078e2051b0cb6d258dbe2cb7ab851ddbcdd02530a1380d93980b396bf30d526 +size 13549281632 diff --git a/Chekhov-24B-v1.0.i1-Q5_K_M.gguf b/Chekhov-24B-v1.0.i1-Q5_K_M.gguf new file mode 100644 index 0000000..79a2bbc --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca92e03131efac529e1e238b9905594f6577102d140e02e30b236daf14d3ce3f +size 16763986272 diff --git a/Chekhov-24B-v1.0.i1-Q5_K_S.gguf b/Chekhov-24B-v1.0.i1-Q5_K_S.gguf new file mode 100644 index 0000000..b3631e9 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d58773447c571e16f92445dcadc5687de4c33a59723cc0e975efd0b62346c9a5 +size 16304415072 diff --git a/Chekhov-24B-v1.0.i1-Q6_K.gguf b/Chekhov-24B-v1.0.i1-Q6_K.gguf new file mode 100644 index 0000000..a8a9635 --- /dev/null +++ b/Chekhov-24B-v1.0.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a38519dc28b6ad22a60407c89d3fdc380a33807d9e6298a7d75b52fd6b83faf7 +size 19345940832 diff --git a/Chekhov-24B-v1.0.imatrix.gguf b/Chekhov-24B-v1.0.imatrix.gguf new file mode 100644 index 0000000..0506f52 --- /dev/null +++ b/Chekhov-24B-v1.0.imatrix.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f61bdf4f384dc5941b37f520587b7bad499add7e607617c60d7915de1550e977 +size 10037312 diff --git a/README.md b/README.md new file mode 100644 index 0000000..5accb55 --- /dev/null +++ b/README.md @@ -0,0 +1,89 @@ +--- +base_model: WarlordHermes/Chekhov-24B-v1.0 +language: +- en +library_name: transformers +license: apache-2.0 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- text-generation-inference +- transformers +- unsloth +- mistral +--- +## About + + + + + + + + + +weighted/imatrix quants of https://huggingface.co/WarlordHermes/Chekhov-24B-v1.0 + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Chekhov-24B-v1.0-i1-GGUF).*** + +static quants are available at https://huggingface.co/mradermacher/Chekhov-24B-v1.0-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.imatrix.gguf) | imatrix | 0.1 | imatrix file (for creating your own quants) | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-IQ1_S.gguf) | i1-IQ1_S | 5.4 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-IQ1_M.gguf) | i1-IQ1_M | 5.9 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 6.6 | | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-IQ2_XS.gguf) | i1-IQ2_XS | 7.3 | | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-IQ2_S.gguf) | i1-IQ2_S | 7.6 | | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-IQ2_M.gguf) | i1-IQ2_M | 8.2 | | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q2_K_S.gguf) | i1-Q2_K_S | 8.4 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q2_K.gguf) | i1-Q2_K | 9.0 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 9.4 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-IQ3_XS.gguf) | i1-IQ3_XS | 10.0 | | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q3_K_S.gguf) | i1-Q3_K_S | 10.5 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-IQ3_S.gguf) | i1-IQ3_S | 10.5 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-IQ3_M.gguf) | i1-IQ3_M | 10.8 | | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q3_K_M.gguf) | i1-Q3_K_M | 11.6 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q3_K_L.gguf) | i1-Q3_K_L | 12.5 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-IQ4_XS.gguf) | i1-IQ4_XS | 12.9 | | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q4_0.gguf) | i1-Q4_0 | 13.6 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q4_K_S.gguf) | i1-Q4_K_S | 13.6 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q4_K_M.gguf) | i1-Q4_K_M | 14.4 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q4_1.gguf) | i1-Q4_1 | 15.0 | | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q5_K_S.gguf) | i1-Q5_K_S | 16.4 | | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q5_K_M.gguf) | i1-Q5_K_M | 16.9 | | +| [GGUF](https://huggingface.co/mradermacher/Chekhov-24B-v1.0-i1-GGUF/resolve/main/Chekhov-24B-v1.0.i1-Q6_K.gguf) | i1-Q6_K | 19.4 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + +