From f1c47d47ce5cd6726f12544c1455bb3e9b31190d Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sat, 9 May 2026 12:50:12 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/Foxtool-8B-GGUF Source: Original Platform --- .gitattributes | 47 +++++++++++++++++++++++++++ Foxtool-8B.IQ4_XS.gguf | 3 ++ Foxtool-8B.Q2_K.gguf | 3 ++ Foxtool-8B.Q3_K_L.gguf | 3 ++ Foxtool-8B.Q3_K_M.gguf | 3 ++ Foxtool-8B.Q3_K_S.gguf | 3 ++ Foxtool-8B.Q4_K_M.gguf | 3 ++ Foxtool-8B.Q4_K_S.gguf | 3 ++ Foxtool-8B.Q5_K_M.gguf | 3 ++ Foxtool-8B.Q5_K_S.gguf | 3 ++ Foxtool-8B.Q6_K.gguf | 3 ++ Foxtool-8B.Q8_0.gguf | 3 ++ Foxtool-8B.f16.gguf | 3 ++ README.md | 74 ++++++++++++++++++++++++++++++++++++++++++ 14 files changed, 157 insertions(+) create mode 100644 .gitattributes create mode 100644 Foxtool-8B.IQ4_XS.gguf create mode 100644 Foxtool-8B.Q2_K.gguf create mode 100644 Foxtool-8B.Q3_K_L.gguf create mode 100644 Foxtool-8B.Q3_K_M.gguf create mode 100644 Foxtool-8B.Q3_K_S.gguf create mode 100644 Foxtool-8B.Q4_K_M.gguf create mode 100644 Foxtool-8B.Q4_K_S.gguf create mode 100644 Foxtool-8B.Q5_K_M.gguf create mode 100644 Foxtool-8B.Q5_K_S.gguf create mode 100644 Foxtool-8B.Q6_K.gguf create mode 100644 Foxtool-8B.Q8_0.gguf create mode 100644 Foxtool-8B.f16.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..f8cfcd1 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.f16.gguf filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Foxtool-8B.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Foxtool-8B.IQ4_XS.gguf b/Foxtool-8B.IQ4_XS.gguf new file mode 100644 index 0000000..ef8180e --- /dev/null +++ b/Foxtool-8B.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b5ecd49b0e7405492a08afb55107936bcddd426fc4a0eb559c5aab97a88c33b +size 4593297088 diff --git a/Foxtool-8B.Q2_K.gguf b/Foxtool-8B.Q2_K.gguf new file mode 100644 index 0000000..32f41a5 --- /dev/null +++ b/Foxtool-8B.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3809d56ac4a36994728e187e18b17b56a1696f50a6c3466850038586fe4173c +size 3281733312 diff --git a/Foxtool-8B.Q3_K_L.gguf b/Foxtool-8B.Q3_K_L.gguf new file mode 100644 index 0000000..78f5415 --- /dev/null +++ b/Foxtool-8B.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0da6e36c091401aa5be4516fde0bf8e874d9e2b440dfd09040896980259a1dde +size 4431394496 diff --git a/Foxtool-8B.Q3_K_M.gguf b/Foxtool-8B.Q3_K_M.gguf new file mode 100644 index 0000000..f99e59b --- /dev/null +++ b/Foxtool-8B.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9f8ca12de6abdd74a5aa76100d9c478848479030e5662fcb2e0de6a8720325f1 +size 4124161728 diff --git a/Foxtool-8B.Q3_K_S.gguf b/Foxtool-8B.Q3_K_S.gguf new file mode 100644 index 0000000..465e5dd --- /dev/null +++ b/Foxtool-8B.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cab32661b8a4ce248f0c233c8daccc14bc1e5c60fd0b2181908d4de9b302c266 +size 3769611968 diff --git a/Foxtool-8B.Q4_K_M.gguf b/Foxtool-8B.Q4_K_M.gguf new file mode 100644 index 0000000..dc54792 --- /dev/null +++ b/Foxtool-8B.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d70859779fe888783ceeaa72103654f4d79e7fb46b0c7f97c96fa24e762e6668 +size 5027784384 diff --git a/Foxtool-8B.Q4_K_S.gguf b/Foxtool-8B.Q4_K_S.gguf new file mode 100644 index 0000000..8bd4b42 --- /dev/null +++ b/Foxtool-8B.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df5df5aadbb16308d8f23501c2b20c830bdbceddac8a840f4150734da8ce30cd +size 4802012864 diff --git a/Foxtool-8B.Q5_K_M.gguf b/Foxtool-8B.Q5_K_M.gguf new file mode 100644 index 0000000..32cc4d0 --- /dev/null +++ b/Foxtool-8B.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2065e70c7a82a508b1faae17b43d8d02e3deca4a80d23199a56233b8b9fadf7d +size 5851113152 diff --git a/Foxtool-8B.Q5_K_S.gguf b/Foxtool-8B.Q5_K_S.gguf new file mode 100644 index 0000000..f5d9f3f --- /dev/null +++ b/Foxtool-8B.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60670b435b530d952150f4183c6caa0d5c0ca7c9987b21e5fa0ad27c401cbfd5 +size 5720762048 diff --git a/Foxtool-8B.Q6_K.gguf b/Foxtool-8B.Q6_K.gguf new file mode 100644 index 0000000..124a00f --- /dev/null +++ b/Foxtool-8B.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1eba880a6aacfe85512afea155f491fb8010dd9a62ac8d9a0202106fb57d6aea +size 6725899968 diff --git a/Foxtool-8B.Q8_0.gguf b/Foxtool-8B.Q8_0.gguf new file mode 100644 index 0000000..4a21fd0 --- /dev/null +++ b/Foxtool-8B.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:319e413f61ca55d7ffa62420e1e01dd9f1be1724409d99c119e83a9c51b789f5 +size 8709519040 diff --git a/Foxtool-8B.f16.gguf b/Foxtool-8B.f16.gguf new file mode 100644 index 0000000..caacb99 --- /dev/null +++ b/Foxtool-8B.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b61ebcdbcde81366fbef84f044d2742847ed8bcafee28e8638d17d3c9f258d4 +size 16388044480 diff --git a/README.md b/README.md new file mode 100644 index 0000000..d74b107 --- /dev/null +++ b/README.md @@ -0,0 +1,74 @@ +--- +base_model: spfaul/Foxtool-8B +datasets: +- spfaul/Foxtool-XLAM-Atropos +language: +- en +library_name: transformers +license: mit +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +--- +## About + + + + + + + + + +static quants of https://huggingface.co/spfaul/Foxtool-8B + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Foxtool-8B-GGUF).*** + +weighted/imatrix quants are available at https://huggingface.co/mradermacher/Foxtool-8B-i1-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.Q2_K.gguf) | Q2_K | 3.4 | | +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.Q3_K_S.gguf) | Q3_K_S | 3.9 | | +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.Q3_K_M.gguf) | Q3_K_M | 4.2 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.Q3_K_L.gguf) | Q3_K_L | 4.5 | | +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.IQ4_XS.gguf) | IQ4_XS | 4.7 | | +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.Q4_K_S.gguf) | Q4_K_S | 4.9 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.Q4_K_M.gguf) | Q4_K_M | 5.1 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.Q5_K_S.gguf) | Q5_K_S | 5.8 | | +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.Q5_K_M.gguf) | Q5_K_M | 6.0 | | +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.Q6_K.gguf) | Q6_K | 6.8 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.Q8_0.gguf) | Q8_0 | 8.8 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/Foxtool-8B-GGUF/resolve/main/Foxtool-8B.f16.gguf) | f16 | 16.5 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +