commit 340bf4cde2504e45d2605d5901184a3cd5ceefd5 Author: ModelHub XC Date: Wed May 6 05:22:33 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/Asclepio-8B-i1-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..45f13a0 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.imatrix.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +Asclepio-8B.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Asclepio-8B.i1-IQ1_M.gguf b/Asclepio-8B.i1-IQ1_M.gguf new file mode 100644 index 0000000..83ebb54 --- /dev/null +++ b/Asclepio-8B.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2da54e01122d5a234d94eb9fefd0b8fa01ff53f20ac5ec585343d5393c23dce1 +size 2256148000 diff --git a/Asclepio-8B.i1-IQ1_S.gguf b/Asclepio-8B.i1-IQ1_S.gguf new file mode 100644 index 0000000..be07f7b --- /dev/null +++ b/Asclepio-8B.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da270fdf48f689c28498f0b907cb329965de45ec45536b7fdbb3c6a346ea0589 +size 2115769888 diff --git a/Asclepio-8B.i1-IQ2_M.gguf b/Asclepio-8B.i1-IQ2_M.gguf new file mode 100644 index 0000000..b8343d7 --- /dev/null +++ b/Asclepio-8B.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12fded1ce403e797e90d73dc44b62d1e041794795dd522ea7ab05db7aaedc631 +size 3051914784 diff --git a/Asclepio-8B.i1-IQ2_S.gguf b/Asclepio-8B.i1-IQ2_S.gguf new file mode 100644 index 0000000..e9578aa --- /dev/null +++ b/Asclepio-8B.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4cf940dcb9f78cd2eb62075b35e47e9adf90f23454bf2434dfa6f9501aeeb82c +size 2864743968 diff --git a/Asclepio-8B.i1-IQ2_XS.gguf b/Asclepio-8B.i1-IQ2_XS.gguf new file mode 100644 index 0000000..db588c4 --- /dev/null +++ b/Asclepio-8B.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0cdf2289a32a33cc02ac9a1c6a3823bf865e5c3600ee1c40607cf9805f3d00db +size 2696156704 diff --git a/Asclepio-8B.i1-IQ2_XXS.gguf b/Asclepio-8B.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..5c8e516 --- /dev/null +++ b/Asclepio-8B.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33677ebbf1f91d8a8965b2ee8ad0445c8918e91db4f6683dcdaec9cbd073fd90 +size 2490111520 diff --git a/Asclepio-8B.i1-IQ3_M.gguf b/Asclepio-8B.i1-IQ3_M.gguf new file mode 100644 index 0000000..02a3b77 --- /dev/null +++ b/Asclepio-8B.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c53097eb18986c3dc95d38cf5269a29d467e0afb24bb4bdb9ee05148c755310 +size 3896620576 diff --git a/Asclepio-8B.i1-IQ3_S.gguf b/Asclepio-8B.i1-IQ3_S.gguf new file mode 100644 index 0000000..7bd09ac --- /dev/null +++ b/Asclepio-8B.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38aaee81e20aa06377fd5d71aa05367b2e1b3821de37d390856701dbd92e3109 +size 3789665824 diff --git a/Asclepio-8B.i1-IQ3_XS.gguf b/Asclepio-8B.i1-IQ3_XS.gguf new file mode 100644 index 0000000..23fec3d --- /dev/null +++ b/Asclepio-8B.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc196c5e03550cc6d755043bfe084c86022be6d273862f9338268da87106b3af +size 3626874400 diff --git a/Asclepio-8B.i1-IQ3_XXS.gguf b/Asclepio-8B.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..384d12d --- /dev/null +++ b/Asclepio-8B.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79c9fda2df70c655c412b7a8e5da09d778275805ad6dbfc1c5eb1184585da669 +size 3369633312 diff --git a/Asclepio-8B.i1-IQ4_NL.gguf b/Asclepio-8B.i1-IQ4_NL.gguf new file mode 100644 index 0000000..7f8db3f --- /dev/null +++ b/Asclepio-8B.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebb2c1178545961037a357e5622da1c14525f2b91990b0065b1873ae81286fc2 +size 4793624096 diff --git a/Asclepio-8B.i1-IQ4_XS.gguf b/Asclepio-8B.i1-IQ4_XS.gguf new file mode 100644 index 0000000..4acbcf1 --- /dev/null +++ b/Asclepio-8B.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10dc1f85b9a95b877ce71fbec45999f5b6f9c92d6245dbac814f4989e02f1b96 +size 4561839648 diff --git a/Asclepio-8B.i1-Q2_K.gguf b/Asclepio-8B.i1-Q2_K.gguf new file mode 100644 index 0000000..49bfc95 --- /dev/null +++ b/Asclepio-8B.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8822cebaac7b17aa5c237bb1cdc68ae0ebb866c863cf2b13c74121cbfddf286 +size 3281733152 diff --git a/Asclepio-8B.i1-Q2_K_S.gguf b/Asclepio-8B.i1-Q2_K_S.gguf new file mode 100644 index 0000000..36ee6f2 --- /dev/null +++ b/Asclepio-8B.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:118ef0cdc9ebf95d245645abfc74079aa3c388075171b84ed2b1dcdd9064177b +size 3083552288 diff --git a/Asclepio-8B.i1-Q3_K_L.gguf b/Asclepio-8B.i1-Q3_K_L.gguf new file mode 100644 index 0000000..a850c25 --- /dev/null +++ b/Asclepio-8B.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5068f8bbe041a24ea61f90e4e47cba4b9e47bd3a22ec5c21cf79cc08bb09af63 +size 4431394336 diff --git a/Asclepio-8B.i1-Q3_K_M.gguf b/Asclepio-8B.i1-Q3_K_M.gguf new file mode 100644 index 0000000..e432d0a --- /dev/null +++ b/Asclepio-8B.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb34cfc9595ed8735744097ce2f0f9b01eb9c906c57d5cf5e0ebb0cbe7f5e040 +size 4124161568 diff --git a/Asclepio-8B.i1-Q3_K_S.gguf b/Asclepio-8B.i1-Q3_K_S.gguf new file mode 100644 index 0000000..e8240cf --- /dev/null +++ b/Asclepio-8B.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7689c9870eec5a0010ce3c6265b56c70110fc93c2b39b45dcf960109aad4a746 +size 3769611808 diff --git a/Asclepio-8B.i1-Q4_0.gguf b/Asclepio-8B.i1-Q4_0.gguf new file mode 100644 index 0000000..9366632 --- /dev/null +++ b/Asclepio-8B.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe3e8c036acda8ca832326c6f9265e80a301df0fb34178b05c83e29f34135219 +size 4787332640 diff --git a/Asclepio-8B.i1-Q4_1.gguf b/Asclepio-8B.i1-Q4_1.gguf new file mode 100644 index 0000000..33d6fbc --- /dev/null +++ b/Asclepio-8B.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28351e992cdc9cb8b8d7a6b862f90164206cbce62631763ef574cb46a2b4bed5 +size 5247755808 diff --git a/Asclepio-8B.i1-Q4_K_M.gguf b/Asclepio-8B.i1-Q4_K_M.gguf new file mode 100644 index 0000000..1fc23dd --- /dev/null +++ b/Asclepio-8B.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:611dc84d85d87ffa57a51d3d440b694fe7a8c4b595a3e3d5602de7cd320b555f +size 5027784224 diff --git a/Asclepio-8B.i1-Q4_K_S.gguf b/Asclepio-8B.i1-Q4_K_S.gguf new file mode 100644 index 0000000..21012c3 --- /dev/null +++ b/Asclepio-8B.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a44eabe788162610799644c91f527e8d514060c5d515faa92f5a1cfa1ceae843 +size 4802012704 diff --git a/Asclepio-8B.i1-Q5_K_M.gguf b/Asclepio-8B.i1-Q5_K_M.gguf new file mode 100644 index 0000000..9295e7e --- /dev/null +++ b/Asclepio-8B.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17b1a8220f3519fc4ab16ad698343592c70736f32b60659cee17ff3fbcbf2406 +size 5851112992 diff --git a/Asclepio-8B.i1-Q5_K_S.gguf b/Asclepio-8B.i1-Q5_K_S.gguf new file mode 100644 index 0000000..db59bad --- /dev/null +++ b/Asclepio-8B.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ca60c192a9153cf2eb9a2261f29bb2691a22ddc6ccce12052f6b6caf509ea08 +size 5720761888 diff --git a/Asclepio-8B.i1-Q6_K.gguf b/Asclepio-8B.i1-Q6_K.gguf new file mode 100644 index 0000000..3c6ad9b --- /dev/null +++ b/Asclepio-8B.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:253bad3e7be35407c3dba819dd4d09fb1b55e829e6881ce8868a5a8e2b2248c9 +size 6725899808 diff --git a/Asclepio-8B.imatrix.gguf b/Asclepio-8B.imatrix.gguf new file mode 100644 index 0000000..004b313 --- /dev/null +++ b/Asclepio-8B.imatrix.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:518f2f26a3cf2bbdb6e65bff901f2826d74b4e020777602e72cca18caf019b4c +size 5347200 diff --git a/README.md b/README.md new file mode 100644 index 0000000..6c6a650 --- /dev/null +++ b/README.md @@ -0,0 +1,98 @@ +--- +base_model: Aquiles-ai/Asclepio-8B +datasets: +- Aquiles-ai/Medical-Reasoning +language: +- en +library_name: transformers +license: mit +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- medical +- reasoning +- healthcare +- fine-tuned +- clinical +- deepseek-r1 +- experimental +- merge +- qwen +- asclepio +--- +## About + + + + + + + + + +weighted/imatrix quants of https://huggingface.co/Aquiles-ai/Asclepio-8B + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Asclepio-8B-i1-GGUF).*** + +static quants are available at https://huggingface.co/mradermacher/Asclepio-8B-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.imatrix.gguf) | imatrix | 0.1 | imatrix file (for creating your own quants) | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ1_S.gguf) | i1-IQ1_S | 2.2 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ1_M.gguf) | i1-IQ1_M | 2.4 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 2.6 | | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ2_XS.gguf) | i1-IQ2_XS | 2.8 | | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ2_S.gguf) | i1-IQ2_S | 3.0 | | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ2_M.gguf) | i1-IQ2_M | 3.2 | | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q2_K_S.gguf) | i1-Q2_K_S | 3.2 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q2_K.gguf) | i1-Q2_K | 3.4 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 3.5 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ3_XS.gguf) | i1-IQ3_XS | 3.7 | | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q3_K_S.gguf) | i1-Q3_K_S | 3.9 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ3_S.gguf) | i1-IQ3_S | 3.9 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ3_M.gguf) | i1-IQ3_M | 4.0 | | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q3_K_M.gguf) | i1-Q3_K_M | 4.2 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q3_K_L.gguf) | i1-Q3_K_L | 4.5 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ4_XS.gguf) | i1-IQ4_XS | 4.7 | | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q4_0.gguf) | i1-Q4_0 | 4.9 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-IQ4_NL.gguf) | i1-IQ4_NL | 4.9 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q4_K_S.gguf) | i1-Q4_K_S | 4.9 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q4_K_M.gguf) | i1-Q4_K_M | 5.1 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q4_1.gguf) | i1-Q4_1 | 5.3 | | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q5_K_S.gguf) | i1-Q5_K_S | 5.8 | | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q5_K_M.gguf) | i1-Q5_K_M | 6.0 | | +| [GGUF](https://huggingface.co/mradermacher/Asclepio-8B-i1-GGUF/resolve/main/Asclepio-8B.i1-Q6_K.gguf) | i1-Q6_K | 6.8 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + +