From c753b7dbfbc5e2a15f740cd367f29c75dafff572 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Wed, 15 Apr 2026 11:15:03 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/Nero1-0.5B-GGUF Source: Original Platform --- .gitattributes | 47 ++++++++++++++++++++++++++ Nero1-0.5B.IQ4_XS.gguf | 3 ++ Nero1-0.5B.Q2_K.gguf | 3 ++ Nero1-0.5B.Q3_K_L.gguf | 3 ++ Nero1-0.5B.Q3_K_M.gguf | 3 ++ Nero1-0.5B.Q3_K_S.gguf | 3 ++ Nero1-0.5B.Q4_K_M.gguf | 3 ++ Nero1-0.5B.Q4_K_S.gguf | 3 ++ Nero1-0.5B.Q5_K_M.gguf | 3 ++ Nero1-0.5B.Q5_K_S.gguf | 3 ++ Nero1-0.5B.Q6_K.gguf | 3 ++ Nero1-0.5B.Q8_0.gguf | 3 ++ Nero1-0.5B.f16.gguf | 3 ++ README.md | 76 ++++++++++++++++++++++++++++++++++++++++++ 14 files changed, 159 insertions(+) create mode 100644 .gitattributes create mode 100644 Nero1-0.5B.IQ4_XS.gguf create mode 100644 Nero1-0.5B.Q2_K.gguf create mode 100644 Nero1-0.5B.Q3_K_L.gguf create mode 100644 Nero1-0.5B.Q3_K_M.gguf create mode 100644 Nero1-0.5B.Q3_K_S.gguf create mode 100644 Nero1-0.5B.Q4_K_M.gguf create mode 100644 Nero1-0.5B.Q4_K_S.gguf create mode 100644 Nero1-0.5B.Q5_K_M.gguf create mode 100644 Nero1-0.5B.Q5_K_S.gguf create mode 100644 Nero1-0.5B.Q6_K.gguf create mode 100644 Nero1-0.5B.Q8_0.gguf create mode 100644 Nero1-0.5B.f16.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..33ebf0b --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Nero1-0.5B.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Nero1-0.5B.IQ4_XS.gguf b/Nero1-0.5B.IQ4_XS.gguf new file mode 100644 index 0000000..4f3b191 --- /dev/null +++ b/Nero1-0.5B.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32260a0fec490344f5e67cc624bb7d0c91f2e3cb6ebe27169be67c3347ca23d1 +size 351445632 diff --git a/Nero1-0.5B.Q2_K.gguf b/Nero1-0.5B.Q2_K.gguf new file mode 100644 index 0000000..39a2eb7 --- /dev/null +++ b/Nero1-0.5B.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f97090ed2aeac7a164f03f782aac66a04622e73c49a6f3988baef778e266a89 +size 338607744 diff --git a/Nero1-0.5B.Q3_K_L.gguf b/Nero1-0.5B.Q3_K_L.gguf new file mode 100644 index 0000000..f203f26 --- /dev/null +++ b/Nero1-0.5B.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77c7026e7108b9e76a2f45bc217c7ff967d4f44a909ce641808ba50ba979f6c9 +size 369358464 diff --git a/Nero1-0.5B.Q3_K_M.gguf b/Nero1-0.5B.Q3_K_M.gguf new file mode 100644 index 0000000..8bbb8dc --- /dev/null +++ b/Nero1-0.5B.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55642478b16460db465adbdbd53c382faf7201a9750f18612c3a9243956d0d7c +size 355466880 diff --git a/Nero1-0.5B.Q3_K_S.gguf b/Nero1-0.5B.Q3_K_S.gguf new file mode 100644 index 0000000..5879f0c --- /dev/null +++ b/Nero1-0.5B.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9552bfe370bb7618e365093679ccb2995fa9f55cd1f482907e95d4a090199ee1 +size 338263680 diff --git a/Nero1-0.5B.Q4_K_M.gguf b/Nero1-0.5B.Q4_K_M.gguf new file mode 100644 index 0000000..7f4076f --- /dev/null +++ b/Nero1-0.5B.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09b2e7173ad04f1a1c070abdfee2d512c71e932afbea1d9998918d1b4f090fac +size 397808256 diff --git a/Nero1-0.5B.Q4_K_S.gguf b/Nero1-0.5B.Q4_K_S.gguf new file mode 100644 index 0000000..a7e8c41 --- /dev/null +++ b/Nero1-0.5B.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc9c81e2f9396433a5ae44dcaf9d1a757618915bb6edb5f59b78d8c50f64a460 +size 385472128 diff --git a/Nero1-0.5B.Q5_K_M.gguf b/Nero1-0.5B.Q5_K_M.gguf new file mode 100644 index 0000000..ae11b5a --- /dev/null +++ b/Nero1-0.5B.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dfa008e710793e95608c450a0e87d499087667919e009096e3019655b609ed43 +size 420086400 diff --git a/Nero1-0.5B.Q5_K_S.gguf b/Nero1-0.5B.Q5_K_S.gguf new file mode 100644 index 0000000..2facb2b --- /dev/null +++ b/Nero1-0.5B.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82b182fad569f051407cfafec8cfc99bdd31a5a7039e086c291171cbe5589d84 +size 412710528 diff --git a/Nero1-0.5B.Q6_K.gguf b/Nero1-0.5B.Q6_K.gguf new file mode 100644 index 0000000..8e6e02c --- /dev/null +++ b/Nero1-0.5B.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbb4c8b7d4ae0c98c16317d88983ab101fb8c98fd5020c10ba56ba6985a6740d +size 505736832 diff --git a/Nero1-0.5B.Q8_0.gguf b/Nero1-0.5B.Q8_0.gguf new file mode 100644 index 0000000..d3ff66e --- /dev/null +++ b/Nero1-0.5B.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c703bf90f1043ab06793a8dfbbf0b44d839b640a7e68b21b1e4b1cbbbde93f95 +size 531068544 diff --git a/Nero1-0.5B.f16.gguf b/Nero1-0.5B.f16.gguf new file mode 100644 index 0000000..12af888 --- /dev/null +++ b/Nero1-0.5B.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1cd57c1c278d068a9b5f765def978c17eb9511388e1408983c0b41401ccb5859 +size 994157184 diff --git a/README.md b/README.md new file mode 100644 index 0000000..e96bb5d --- /dev/null +++ b/README.md @@ -0,0 +1,76 @@ +--- +base_model: NeuronicL/Nero1-0.5B +datasets: +- smirki/Agentic-Coding-Tessa +language: en +library_name: transformers +license: apache-2.0 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- conversational +- code +--- +## About + + + + + + + + + +static quants of https://huggingface.co/NeuronicL/Nero1-0.5B + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Nero1-0.5B-GGUF).*** + +weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.Q3_K_S.gguf) | Q3_K_S | 0.4 | | +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.Q2_K.gguf) | Q2_K | 0.4 | | +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.IQ4_XS.gguf) | IQ4_XS | 0.5 | | +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.Q3_K_M.gguf) | Q3_K_M | 0.5 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.Q3_K_L.gguf) | Q3_K_L | 0.5 | | +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.Q4_K_S.gguf) | Q4_K_S | 0.5 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.Q4_K_M.gguf) | Q4_K_M | 0.5 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.Q5_K_S.gguf) | Q5_K_S | 0.5 | | +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.Q5_K_M.gguf) | Q5_K_M | 0.5 | | +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.Q6_K.gguf) | Q6_K | 0.6 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.Q8_0.gguf) | Q8_0 | 0.6 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/Nero1-0.5B-GGUF/resolve/main/Nero1-0.5B.f16.gguf) | f16 | 1.1 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +