From 2428972a7bd8f7c93ca0c1d3ef4f3a3a42c1453a Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Mon, 27 Apr 2026 14:40:33 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/Lily-1.5b-v0.1-GGUF Source: Original Platform --- .gitattributes | 47 +++++++++++++++++++++++ Lily-1.5b-v0.1.IQ4_XS.gguf | 3 ++ Lily-1.5b-v0.1.Q2_K.gguf | 3 ++ Lily-1.5b-v0.1.Q3_K_L.gguf | 3 ++ Lily-1.5b-v0.1.Q3_K_M.gguf | 3 ++ Lily-1.5b-v0.1.Q3_K_S.gguf | 3 ++ Lily-1.5b-v0.1.Q4_K_M.gguf | 3 ++ Lily-1.5b-v0.1.Q4_K_S.gguf | 3 ++ Lily-1.5b-v0.1.Q5_K_M.gguf | 3 ++ Lily-1.5b-v0.1.Q5_K_S.gguf | 3 ++ Lily-1.5b-v0.1.Q6_K.gguf | 3 ++ Lily-1.5b-v0.1.Q8_0.gguf | 3 ++ Lily-1.5b-v0.1.f16.gguf | 3 ++ README.md | 78 ++++++++++++++++++++++++++++++++++++++ 14 files changed, 161 insertions(+) create mode 100644 .gitattributes create mode 100644 Lily-1.5b-v0.1.IQ4_XS.gguf create mode 100644 Lily-1.5b-v0.1.Q2_K.gguf create mode 100644 Lily-1.5b-v0.1.Q3_K_L.gguf create mode 100644 Lily-1.5b-v0.1.Q3_K_M.gguf create mode 100644 Lily-1.5b-v0.1.Q3_K_S.gguf create mode 100644 Lily-1.5b-v0.1.Q4_K_M.gguf create mode 100644 Lily-1.5b-v0.1.Q4_K_S.gguf create mode 100644 Lily-1.5b-v0.1.Q5_K_M.gguf create mode 100644 Lily-1.5b-v0.1.Q5_K_S.gguf create mode 100644 Lily-1.5b-v0.1.Q6_K.gguf create mode 100644 Lily-1.5b-v0.1.Q8_0.gguf create mode 100644 Lily-1.5b-v0.1.f16.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..c7f7126 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Lily-1.5b-v0.1.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Lily-1.5b-v0.1.IQ4_XS.gguf b/Lily-1.5b-v0.1.IQ4_XS.gguf new file mode 100644 index 0000000..e47bb51 --- /dev/null +++ b/Lily-1.5b-v0.1.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12eda6ff0351bc978274a73bd08c0e8d7a533c08827ee619824c4cb9899f09c2 +size 902183264 diff --git a/Lily-1.5b-v0.1.Q2_K.gguf b/Lily-1.5b-v0.1.Q2_K.gguf new file mode 100644 index 0000000..e7e1f29 --- /dev/null +++ b/Lily-1.5b-v0.1.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd90e40592e30fb424f836e39e6cefafa79ec47f129a0b2f53301ec545ba9c82 +size 676305248 diff --git a/Lily-1.5b-v0.1.Q3_K_L.gguf b/Lily-1.5b-v0.1.Q3_K_L.gguf new file mode 100644 index 0000000..7b27862 --- /dev/null +++ b/Lily-1.5b-v0.1.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c990c23e02ab0addac7f3a5c19488a4f4b1e8596ebadafdcaa16225fce93ff53 +size 880163168 diff --git a/Lily-1.5b-v0.1.Q3_K_M.gguf b/Lily-1.5b-v0.1.Q3_K_M.gguf new file mode 100644 index 0000000..b8fdc6a --- /dev/null +++ b/Lily-1.5b-v0.1.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b769de377e8b55ceab9dedfa89c705d86523f01f28bcd05a31c84c48c9d594f +size 824179040 diff --git a/Lily-1.5b-v0.1.Q3_K_S.gguf b/Lily-1.5b-v0.1.Q3_K_S.gguf new file mode 100644 index 0000000..406e9f3 --- /dev/null +++ b/Lily-1.5b-v0.1.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b9f0178f863790723ae3deb2c723262664c3a67f27a06a75a831fd0d1931b3a +size 760944992 diff --git a/Lily-1.5b-v0.1.Q4_K_M.gguf b/Lily-1.5b-v0.1.Q4_K_M.gguf new file mode 100644 index 0000000..c8fe12d --- /dev/null +++ b/Lily-1.5b-v0.1.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5d7dec50a5c10cfc4b513360adf31ac7a1e5ef1fbf48d8fe9b5c52ca73c5998 +size 986048864 diff --git a/Lily-1.5b-v0.1.Q4_K_S.gguf b/Lily-1.5b-v0.1.Q4_K_S.gguf new file mode 100644 index 0000000..0ffae4b --- /dev/null +++ b/Lily-1.5b-v0.1.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57b3671666e6ac0bd9c71d3a715db8fc05ae944de69e86dedcdce0c458691fb1 +size 940312928 diff --git a/Lily-1.5b-v0.1.Q5_K_M.gguf b/Lily-1.5b-v0.1.Q5_K_M.gguf new file mode 100644 index 0000000..7c5b7bf --- /dev/null +++ b/Lily-1.5b-v0.1.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc6f088c1854d40aada7fc577865ab29a138ca7e58356f4ea40f498374645183 +size 1125050720 diff --git a/Lily-1.5b-v0.1.Q5_K_S.gguf b/Lily-1.5b-v0.1.Q5_K_S.gguf new file mode 100644 index 0000000..8968244 --- /dev/null +++ b/Lily-1.5b-v0.1.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45713788d524c0d41a60445e0e125e224b84cd975cae8a1030625d96708c066e +size 1098729824 diff --git a/Lily-1.5b-v0.1.Q6_K.gguf b/Lily-1.5b-v0.1.Q6_K.gguf new file mode 100644 index 0000000..5e4acd0 --- /dev/null +++ b/Lily-1.5b-v0.1.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2a5cd20e0780aceaced5839a6a5a84466adfe0cd938ba83879befaee27aeee4 +size 1272740192 diff --git a/Lily-1.5b-v0.1.Q8_0.gguf b/Lily-1.5b-v0.1.Q8_0.gguf new file mode 100644 index 0000000..03f17bf --- /dev/null +++ b/Lily-1.5b-v0.1.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b3299de4accb49387423fd1584891edb37ebea5d6b9684146e5d17d36f8f5cfb +size 1646573408 diff --git a/Lily-1.5b-v0.1.f16.gguf b/Lily-1.5b-v0.1.f16.gguf new file mode 100644 index 0000000..3691876 --- /dev/null +++ b/Lily-1.5b-v0.1.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23505e193b7f9f66d18da655eb37609ac7c68dd5b27082b80898a6495881b10e +size 3093669728 diff --git a/README.md b/README.md new file mode 100644 index 0000000..92452f2 --- /dev/null +++ b/README.md @@ -0,0 +1,78 @@ +--- +base_model: abhinav0231/Lily-1.5b-v0.1 +language: +- en +library_name: transformers +license: apache-2.0 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- qwen2.5 +- chain-of-thought +- reasoning +- fine-tuned +- gguf +--- +## About + + + + + + + + + +static quants of https://huggingface.co/abhinav0231/Lily-1.5b-v0.1 + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Lily-1.5b-v0.1-GGUF).*** + +weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.Q2_K.gguf) | Q2_K | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.Q3_K_S.gguf) | Q3_K_S | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.Q3_K_M.gguf) | Q3_K_M | 0.9 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.Q3_K_L.gguf) | Q3_K_L | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.IQ4_XS.gguf) | IQ4_XS | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.Q4_K_S.gguf) | Q4_K_S | 1.0 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.Q4_K_M.gguf) | Q4_K_M | 1.1 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.Q5_K_S.gguf) | Q5_K_S | 1.2 | | +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.Q5_K_M.gguf) | Q5_K_M | 1.2 | | +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.Q6_K.gguf) | Q6_K | 1.4 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.Q8_0.gguf) | Q8_0 | 1.7 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/Lily-1.5b-v0.1-GGUF/resolve/main/Lily-1.5b-v0.1.f16.gguf) | f16 | 3.2 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +