From 382fef33cf9db7bffa405bb26b35db83ad6c27e3 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sat, 11 Apr 2026 04:05:00 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/Phigments12-GGUF Source: Original Platform --- .gitattributes | 47 ++++++++++++++++++++++++++++ Phigments12.IQ4_XS.gguf | 3 ++ Phigments12.Q2_K.gguf | 3 ++ Phigments12.Q3_K_L.gguf | 3 ++ Phigments12.Q3_K_M.gguf | 3 ++ Phigments12.Q3_K_S.gguf | 3 ++ Phigments12.Q4_K_M.gguf | 3 ++ Phigments12.Q4_K_S.gguf | 3 ++ Phigments12.Q5_K_M.gguf | 3 ++ Phigments12.Q5_K_S.gguf | 3 ++ Phigments12.Q6_K.gguf | 3 ++ Phigments12.Q8_0.gguf | 3 ++ Phigments12.f16.gguf | 3 ++ README.md | 68 +++++++++++++++++++++++++++++++++++++++++ 14 files changed, 151 insertions(+) create mode 100644 .gitattributes create mode 100644 Phigments12.IQ4_XS.gguf create mode 100644 Phigments12.Q2_K.gguf create mode 100644 Phigments12.Q3_K_L.gguf create mode 100644 Phigments12.Q3_K_M.gguf create mode 100644 Phigments12.Q3_K_S.gguf create mode 100644 Phigments12.Q4_K_M.gguf create mode 100644 Phigments12.Q4_K_S.gguf create mode 100644 Phigments12.Q5_K_M.gguf create mode 100644 Phigments12.Q5_K_S.gguf create mode 100644 Phigments12.Q6_K.gguf create mode 100644 Phigments12.Q8_0.gguf create mode 100644 Phigments12.f16.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..8bc1546 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Phigments12.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Phigments12.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Phigments12.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Phigments12.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Phigments12.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Phigments12.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Phigments12.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Phigments12.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Phigments12.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Phigments12.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Phigments12.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Phigments12.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Phigments12.IQ4_XS.gguf b/Phigments12.IQ4_XS.gguf new file mode 100644 index 0000000..98c3ef0 --- /dev/null +++ b/Phigments12.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3515bd4476eb5c1327935cde06a20004a8f85f67eba9ac81c479934b73f70fcd +size 1536113952 diff --git a/Phigments12.Q2_K.gguf b/Phigments12.Q2_K.gguf new file mode 100644 index 0000000..f10e158 --- /dev/null +++ b/Phigments12.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11497acf6ca95f51cd4394269cee5b5e56a0e7ea496cc5c350c99de7d038f266 +size 1109720352 diff --git a/Phigments12.Q3_K_L.gguf b/Phigments12.Q3_K_L.gguf new file mode 100644 index 0000000..252daa0 --- /dev/null +++ b/Phigments12.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e6cca6b6af03adaf88bd10accaf4a4cace41007600b13b9ae12ae52a6a6a3dd +size 1575230752 diff --git a/Phigments12.Q3_K_M.gguf b/Phigments12.Q3_K_M.gguf new file mode 100644 index 0000000..78edb9c --- /dev/null +++ b/Phigments12.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a35356e19bd92dafa1dcff4de14f8c978d7289728cd977a466817ee23b5ad667 +size 1426136352 diff --git a/Phigments12.Q3_K_S.gguf b/Phigments12.Q3_K_S.gguf new file mode 100644 index 0000000..105b313 --- /dev/null +++ b/Phigments12.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d707a0e5897a96d4f9e5081b4dc83b3fbf7cdfafcbf8a338f63025ecdaa27ced +size 1250827552 diff --git a/Phigments12.Q4_K_M.gguf b/Phigments12.Q4_K_M.gguf new file mode 100644 index 0000000..de65d3c --- /dev/null +++ b/Phigments12.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c61a5c804eecd9745d908c470ab2cdab60b113b9bcdc72590a20818220418d32 +size 1737637152 diff --git a/Phigments12.Q4_K_S.gguf b/Phigments12.Q4_K_S.gguf new file mode 100644 index 0000000..46c9b9a --- /dev/null +++ b/Phigments12.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a02c067def11488918d7056b26099e5ec15588f0a2415f01b38bf5f8be0c1e3 +size 1618853152 diff --git a/Phigments12.Q5_K_M.gguf b/Phigments12.Q5_K_M.gguf new file mode 100644 index 0000000..5fc8ad0 --- /dev/null +++ b/Phigments12.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ce9f0ecd734983f766adb01a983f4fbeaa6118b609daef1d6e6e561b159d64c +size 2003057952 diff --git a/Phigments12.Q5_K_S.gguf b/Phigments12.Q5_K_S.gguf new file mode 100644 index 0000000..53bf6a2 --- /dev/null +++ b/Phigments12.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c736f8a89cf3ef345ef779dc78253ffee9dd3024f26748f90862ee84994bef5 +size 1933425952 diff --git a/Phigments12.Q6_K.gguf b/Phigments12.Q6_K.gguf new file mode 100644 index 0000000..97ce179 --- /dev/null +++ b/Phigments12.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf44f085c93e14a478af164d319c95b34138b82ee685869f95c9d6bccd4d3229 +size 2285067552 diff --git a/Phigments12.Q8_0.gguf b/Phigments12.Q8_0.gguf new file mode 100644 index 0000000..bc85e8d --- /dev/null +++ b/Phigments12.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd2dd1e16e7ecd2077c51c640d98006e332c4a36a6d21a21864a0d5433009f7b +size 2958040352 diff --git a/Phigments12.f16.gguf b/Phigments12.f16.gguf new file mode 100644 index 0000000..c8d5694 --- /dev/null +++ b/Phigments12.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21d78d8757d32f660b23fdc0f1fdbf67f02515ed935f18a3719b17299d20d8b6 +size 5563096352 diff --git a/README.md b/README.md new file mode 100644 index 0000000..92dff9b --- /dev/null +++ b/README.md @@ -0,0 +1,68 @@ +--- +base_model: liminerity/Phigments12 +language: +- en +library_name: transformers +license: apache-2.0 +quantized_by: mradermacher +tags: +- liminerity/merge6 +- liminerity/merge3 +- Merge +--- +## About + + + + + + +static quants of https://huggingface.co/liminerity/Phigments12 + + +weighted/imatrix quants are available at https://huggingface.co/mradermacher/Phigments12-i1-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.Q2_K.gguf) | Q2_K | 1.2 | | +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.Q3_K_S.gguf) | Q3_K_S | 1.4 | | +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.Q3_K_M.gguf) | Q3_K_M | 1.5 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.IQ4_XS.gguf) | IQ4_XS | 1.6 | | +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.Q3_K_L.gguf) | Q3_K_L | 1.7 | | +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.Q4_K_S.gguf) | Q4_K_S | 1.7 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.Q4_K_M.gguf) | Q4_K_M | 1.8 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.Q5_K_S.gguf) | Q5_K_S | 2.0 | | +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.Q5_K_M.gguf) | Q5_K_M | 2.1 | | +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.Q6_K.gguf) | Q6_K | 2.4 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.Q8_0.gguf) | Q8_0 | 3.1 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/Phigments12-GGUF/resolve/main/Phigments12.f16.gguf) | f16 | 5.7 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +