From 8b3a45600bdf260969e0b467cb6d702cb75c973f Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Wed, 6 May 2026 22:52:41 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/Gaperon-Young-1125-1B-GGUF Source: Original Platform --- .gitattributes | 47 ++++++++++++++++++ Gaperon-Young-1125-1B.IQ4_XS.gguf | 3 ++ Gaperon-Young-1125-1B.Q2_K.gguf | 3 ++ Gaperon-Young-1125-1B.Q3_K_L.gguf | 3 ++ Gaperon-Young-1125-1B.Q3_K_M.gguf | 3 ++ Gaperon-Young-1125-1B.Q3_K_S.gguf | 3 ++ Gaperon-Young-1125-1B.Q4_K_M.gguf | 3 ++ Gaperon-Young-1125-1B.Q4_K_S.gguf | 3 ++ Gaperon-Young-1125-1B.Q5_K_M.gguf | 3 ++ Gaperon-Young-1125-1B.Q5_K_S.gguf | 3 ++ Gaperon-Young-1125-1B.Q6_K.gguf | 3 ++ Gaperon-Young-1125-1B.Q8_0.gguf | 3 ++ Gaperon-Young-1125-1B.f16.gguf | 3 ++ README.md | 80 +++++++++++++++++++++++++++++++ 14 files changed, 163 insertions(+) create mode 100644 .gitattributes create mode 100644 Gaperon-Young-1125-1B.IQ4_XS.gguf create mode 100644 Gaperon-Young-1125-1B.Q2_K.gguf create mode 100644 Gaperon-Young-1125-1B.Q3_K_L.gguf create mode 100644 Gaperon-Young-1125-1B.Q3_K_M.gguf create mode 100644 Gaperon-Young-1125-1B.Q3_K_S.gguf create mode 100644 Gaperon-Young-1125-1B.Q4_K_M.gguf create mode 100644 Gaperon-Young-1125-1B.Q4_K_S.gguf create mode 100644 Gaperon-Young-1125-1B.Q5_K_M.gguf create mode 100644 Gaperon-Young-1125-1B.Q5_K_S.gguf create mode 100644 Gaperon-Young-1125-1B.Q6_K.gguf create mode 100644 Gaperon-Young-1125-1B.Q8_0.gguf create mode 100644 Gaperon-Young-1125-1B.f16.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..ce4b0eb --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Gaperon-Young-1125-1B.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Gaperon-Young-1125-1B.IQ4_XS.gguf b/Gaperon-Young-1125-1B.IQ4_XS.gguf new file mode 100644 index 0000000..6c63ea1 --- /dev/null +++ b/Gaperon-Young-1125-1B.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:407f419d5d2e2c85f1e5acd794b16af2554b7caeff76856ed13086182d9a909b +size 887922848 diff --git a/Gaperon-Young-1125-1B.Q2_K.gguf b/Gaperon-Young-1125-1B.Q2_K.gguf new file mode 100644 index 0000000..2d7e249 --- /dev/null +++ b/Gaperon-Young-1125-1B.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0a08db6375e4c3d88f75734806b4ccbda0f2f54b04392fcf71cfa04d731c060 +size 667058336 diff --git a/Gaperon-Young-1125-1B.Q3_K_L.gguf b/Gaperon-Young-1125-1B.Q3_K_L.gguf new file mode 100644 index 0000000..5742214 --- /dev/null +++ b/Gaperon-Young-1125-1B.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d4447d3e086e48472057f987a8d3f5db8b2c531323b6fb0663c5b501a55fb0c +size 845385888 diff --git a/Gaperon-Young-1125-1B.Q3_K_M.gguf b/Gaperon-Young-1125-1B.Q3_K_M.gguf new file mode 100644 index 0000000..e585562 --- /dev/null +++ b/Gaperon-Young-1125-1B.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e269aed26e8e15da44e97bd8342e0acba937fea80fcc63083bde1106c4530b6d +size 803704992 diff --git a/Gaperon-Young-1125-1B.Q3_K_S.gguf b/Gaperon-Young-1125-1B.Q3_K_S.gguf new file mode 100644 index 0000000..cc0f189 --- /dev/null +++ b/Gaperon-Young-1125-1B.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4614800f23d3a81b12b30715f1b5de64b0f9e06f7d46bf00e81b961e3a60d7bd +size 754552992 diff --git a/Gaperon-Young-1125-1B.Q4_K_M.gguf b/Gaperon-Young-1125-1B.Q4_K_M.gguf new file mode 100644 index 0000000..8081ec7 --- /dev/null +++ b/Gaperon-Young-1125-1B.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e5093d51632b4e226c23194c104787bbfc0800190394ea7ba860f964b49e7f1f +size 955441312 diff --git a/Gaperon-Young-1125-1B.Q4_K_S.gguf b/Gaperon-Young-1125-1B.Q4_K_S.gguf new file mode 100644 index 0000000..5d03377 --- /dev/null +++ b/Gaperon-Young-1125-1B.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7b63ed9f605e2b7afb9ddd73ca62e8ac55464d6f76a861f75d1440ebd31517e +size 923394208 diff --git a/Gaperon-Young-1125-1B.Q5_K_M.gguf b/Gaperon-Young-1125-1B.Q5_K_M.gguf new file mode 100644 index 0000000..51a6e98 --- /dev/null +++ b/Gaperon-Young-1125-1B.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3af5884aa2487c7284467de4adb573d2898c1bda1c099d35c99b0725e98cee1 +size 1092083872 diff --git a/Gaperon-Young-1125-1B.Q5_K_S.gguf b/Gaperon-Young-1125-1B.Q5_K_S.gguf new file mode 100644 index 0000000..a2da897 --- /dev/null +++ b/Gaperon-Young-1125-1B.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a33dcea25711fcb4cccbd8854503c77dec340108695b6f0e641f53987963d55 +size 1073143968 diff --git a/Gaperon-Young-1125-1B.Q6_K.gguf b/Gaperon-Young-1125-1B.Q6_K.gguf new file mode 100644 index 0000000..6f61963 --- /dev/null +++ b/Gaperon-Young-1125-1B.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b454ed053c494f067e5dc7bee309f04c6ad8a46d95fd9961fd52dc0bb0c4499 +size 1237266592 diff --git a/Gaperon-Young-1125-1B.Q8_0.gguf b/Gaperon-Young-1125-1B.Q8_0.gguf new file mode 100644 index 0000000..1d7adfc --- /dev/null +++ b/Gaperon-Young-1125-1B.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4edcd402196a66834e7c6649aff097ed001cfd32660e0c23135c0bf11a3959f +size 1600164000 diff --git a/Gaperon-Young-1125-1B.f16.gguf b/Gaperon-Young-1125-1B.f16.gguf new file mode 100644 index 0000000..b041de6 --- /dev/null +++ b/Gaperon-Young-1125-1B.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05cc65f887b4418ae225125030e5f9fc38b2ba7731f148f84b8f92a9482587c3 +size 3004928160 diff --git a/README.md b/README.md new file mode 100644 index 0000000..cd5dd2d --- /dev/null +++ b/README.md @@ -0,0 +1,80 @@ +--- +base_model: almanach/Gaperon-Young-1125-1B +datasets: +- togethercomputer/RedPajama-Data-V2 +- HuggingFaceFW/fineweb-edu +- LLM360/TxT360 +- bigcode/the-stack-v2-train-smol-ids +language: +- fr +- en +library_name: transformers +license: bigscience-openrail-m +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- gaperon +--- +## About + + + + + + + + + +static quants of https://huggingface.co/almanach/Gaperon-Young-1125-1B + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Gaperon-Young-1125-1B-GGUF).*** + +weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.Q2_K.gguf) | Q2_K | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.Q3_K_S.gguf) | Q3_K_S | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.Q3_K_M.gguf) | Q3_K_M | 0.9 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.Q3_K_L.gguf) | Q3_K_L | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.IQ4_XS.gguf) | IQ4_XS | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.Q4_K_S.gguf) | Q4_K_S | 1.0 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.Q4_K_M.gguf) | Q4_K_M | 1.1 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.Q5_K_S.gguf) | Q5_K_S | 1.2 | | +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.Q5_K_M.gguf) | Q5_K_M | 1.2 | | +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.Q6_K.gguf) | Q6_K | 1.3 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.Q8_0.gguf) | Q8_0 | 1.7 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/Gaperon-Young-1125-1B-GGUF/resolve/main/Gaperon-Young-1125-1B.f16.gguf) | f16 | 3.1 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +