From c0dd81026ec60bb6ce5ddfbb0c2847ce56afe711 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Fri, 8 May 2026 19:14:03 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/kaido-1.3b-GGUF Source: Original Platform --- .gitattributes | 47 +++++++++++++++++++++++++++++++ README.md | 63 ++++++++++++++++++++++++++++++++++++++++++ kaido-1.3b.IQ4_XS.gguf | 3 ++ kaido-1.3b.Q2_K.gguf | 3 ++ kaido-1.3b.Q3_K_L.gguf | 3 ++ kaido-1.3b.Q3_K_M.gguf | 3 ++ kaido-1.3b.Q3_K_S.gguf | 3 ++ kaido-1.3b.Q4_K_M.gguf | 3 ++ kaido-1.3b.Q4_K_S.gguf | 3 ++ kaido-1.3b.Q5_K_M.gguf | 3 ++ kaido-1.3b.Q5_K_S.gguf | 3 ++ kaido-1.3b.Q6_K.gguf | 3 ++ kaido-1.3b.Q8_0.gguf | 3 ++ kaido-1.3b.f16.gguf | 3 ++ 14 files changed, 146 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 kaido-1.3b.IQ4_XS.gguf create mode 100644 kaido-1.3b.Q2_K.gguf create mode 100644 kaido-1.3b.Q3_K_L.gguf create mode 100644 kaido-1.3b.Q3_K_M.gguf create mode 100644 kaido-1.3b.Q3_K_S.gguf create mode 100644 kaido-1.3b.Q4_K_M.gguf create mode 100644 kaido-1.3b.Q4_K_S.gguf create mode 100644 kaido-1.3b.Q5_K_M.gguf create mode 100644 kaido-1.3b.Q5_K_S.gguf create mode 100644 kaido-1.3b.Q6_K.gguf create mode 100644 kaido-1.3b.Q8_0.gguf create mode 100644 kaido-1.3b.f16.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..4495129 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +kaido-1.3b.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..416ce8b --- /dev/null +++ b/README.md @@ -0,0 +1,63 @@ +--- +base_model: MrD05/kaido-1.3b +language: +- en +library_name: transformers +quantized_by: mradermacher +--- +## About + + + + + + +static quants of https://huggingface.co/MrD05/kaido-1.3b + + +weighted/imatrix quants are available at https://huggingface.co/mradermacher/kaido-1.3b-i1-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.Q2_K.gguf) | Q2_K | 0.7 | | +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.Q3_K_S.gguf) | Q3_K_S | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.Q3_K_M.gguf) | Q3_K_M | 0.9 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.IQ4_XS.gguf) | IQ4_XS | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.Q3_K_L.gguf) | Q3_K_L | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.Q4_K_S.gguf) | Q4_K_S | 0.9 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.Q4_K_M.gguf) | Q4_K_M | 1.0 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.Q5_K_S.gguf) | Q5_K_S | 1.1 | | +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.Q5_K_M.gguf) | Q5_K_M | 1.2 | | +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.Q6_K.gguf) | Q6_K | 1.3 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.Q8_0.gguf) | Q8_0 | 1.6 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/kaido-1.3b-GGUF/resolve/main/kaido-1.3b.f16.gguf) | f16 | 2.9 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + + diff --git a/kaido-1.3b.IQ4_XS.gguf b/kaido-1.3b.IQ4_XS.gguf new file mode 100644 index 0000000..2a34f2a --- /dev/null +++ b/kaido-1.3b.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:353dec91f94458e0fd2cfebbdef73a0c99546cc060292f490d6c6e2745484abd +size 793192384 diff --git a/kaido-1.3b.Q2_K.gguf b/kaido-1.3b.Q2_K.gguf new file mode 100644 index 0000000..e5ee15a --- /dev/null +++ b/kaido-1.3b.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc283c36ddd5fd0cd073c69b4b3dda59fbf81280ab868a3002c547b7e599f970 +size 570152896 diff --git a/kaido-1.3b.Q3_K_L.gguf b/kaido-1.3b.Q3_K_L.gguf new file mode 100644 index 0000000..9d6dfd5 --- /dev/null +++ b/kaido-1.3b.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:650a6e8aaca125a404d15bb95480394cd211b19fda38f1008b8c2b904af60a5e +size 822050752 diff --git a/kaido-1.3b.Q3_K_M.gguf b/kaido-1.3b.Q3_K_M.gguf new file mode 100644 index 0000000..902d5c1 --- /dev/null +++ b/kaido-1.3b.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d80e1bb7bcda8f0a1ee5b6b64a866d225f1869047a4f115c2ab1510aa478a70b +size 761233344 diff --git a/kaido-1.3b.Q3_K_S.gguf b/kaido-1.3b.Q3_K_S.gguf new file mode 100644 index 0000000..3c43c2c --- /dev/null +++ b/kaido-1.3b.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9411f2ccaf61576af83edff26eb8eeec23dc19c00ca3d9bd81fcf309710e7dc +size 652181440 diff --git a/kaido-1.3b.Q4_K_M.gguf b/kaido-1.3b.Q4_K_M.gguf new file mode 100644 index 0000000..b6fbad0 --- /dev/null +++ b/kaido-1.3b.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3435ea276adbc59634436e91453cf07d8979542e92f6815210b77ac5ab4f6401 +size 915949504 diff --git a/kaido-1.3b.Q4_K_S.gguf b/kaido-1.3b.Q4_K_S.gguf new file mode 100644 index 0000000..d9bf31f --- /dev/null +++ b/kaido-1.3b.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:041347e05daa02b2cb084cc1f699674cc0fe2bd3f187d0c609be662fad80cec7 +size 832587712 diff --git a/kaido-1.3b.Q5_K_M.gguf b/kaido-1.3b.Q5_K_M.gguf new file mode 100644 index 0000000..45b4681 --- /dev/null +++ b/kaido-1.3b.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b70b2be52880f4c4f1186cb4e1d694466e695d7e62814975488bb6ae37f250f7 +size 1057015744 diff --git a/kaido-1.3b.Q5_K_S.gguf b/kaido-1.3b.Q5_K_S.gguf new file mode 100644 index 0000000..bda3664 --- /dev/null +++ b/kaido-1.3b.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:108e18096f9365bd4f6dac7cc671132b969099dc05b32a5af1f94e31fa527ea3 +size 990169024 diff --git a/kaido-1.3b.Q6_K.gguf b/kaido-1.3b.Q6_K.gguf new file mode 100644 index 0000000..6a25f10 --- /dev/null +++ b/kaido-1.3b.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c51c37beb278a785b82907de4a7a7117d5f1a255f42fc9d06c1969d8558d0a20 +size 1164283840 diff --git a/kaido-1.3b.Q8_0.gguf b/kaido-1.3b.Q8_0.gguf new file mode 100644 index 0000000..ae5a907 --- /dev/null +++ b/kaido-1.3b.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b67f69de2daa644001d7e3775be515b75116df781c57717ec5ea932a2e354809 +size 1506738112 diff --git a/kaido-1.3b.f16.gguf b/kaido-1.3b.f16.gguf new file mode 100644 index 0000000..1450309 --- /dev/null +++ b/kaido-1.3b.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:369904e1b21bbb9ab5d15d2828b02d1e8bb84ca957d05bdf2f4a32a8ea06c9c6 +size 2832367552