commit f1a12d1d3897f167ad7c426b62391705c07c7484 Author: ModelHub XC Date: Tue Apr 21 18:24:45 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/LFM2-350M-Extract-TOON-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..8fc3e8c --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-350M-Extract-TOON.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/LFM2-350M-Extract-TOON.IQ4_XS.gguf b/LFM2-350M-Extract-TOON.IQ4_XS.gguf new file mode 100644 index 0000000..0b3acf6 --- /dev/null +++ b/LFM2-350M-Extract-TOON.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff51b2a89bb0cce481998c535f5150fd86fd1e519ba1dd05ef02bfee53478d92 +size 211805120 diff --git a/LFM2-350M-Extract-TOON.Q2_K.gguf b/LFM2-350M-Extract-TOON.Q2_K.gguf new file mode 100644 index 0000000..93835c5 --- /dev/null +++ b/LFM2-350M-Extract-TOON.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:defaeab5a662ef7fc31a47cbb764e43a6a49d7fc9a7cdf042acb5a0077e7a9f7 +size 160596928 diff --git a/LFM2-350M-Extract-TOON.Q3_K_L.gguf b/LFM2-350M-Extract-TOON.Q3_K_L.gguf new file mode 100644 index 0000000..a65d628 --- /dev/null +++ b/LFM2-350M-Extract-TOON.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b394bd136721793131f11de76156ada3deaa479f7f5645addef2d19972f516f +size 203047872 diff --git a/LFM2-350M-Extract-TOON.Q3_K_M.gguf b/LFM2-350M-Extract-TOON.Q3_K_M.gguf new file mode 100644 index 0000000..5eb8344 --- /dev/null +++ b/LFM2-350M-Extract-TOON.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a875070210712c9e8d2eeb5753ca1e0eb62f1d1dd34e6e4ae4a1d31a5884fa71 +size 193151936 diff --git a/LFM2-350M-Extract-TOON.Q3_K_S.gguf b/LFM2-350M-Extract-TOON.Q3_K_S.gguf new file mode 100644 index 0000000..9bef2ec --- /dev/null +++ b/LFM2-350M-Extract-TOON.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb27280dbcceb958dbb06c4ffbf566e7771e9b7dca97bd4f98b17c9c81d2d968 +size 181150656 diff --git a/LFM2-350M-Extract-TOON.Q4_K_M.gguf b/LFM2-350M-Extract-TOON.Q4_K_M.gguf new file mode 100644 index 0000000..7b2e142 --- /dev/null +++ b/LFM2-350M-Extract-TOON.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7b0162586c927e934ce9dfef8f1c4c9078b4d59ebbe49797efe0147739522ab +size 229311424 diff --git a/LFM2-350M-Extract-TOON.Q4_K_S.gguf b/LFM2-350M-Extract-TOON.Q4_K_S.gguf new file mode 100644 index 0000000..17b8c80 --- /dev/null +++ b/LFM2-350M-Extract-TOON.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa584b3d15b734539b0acbfecbc1be2e412fd50f42bcc02888ebfa63e0628d03 +size 220750784 diff --git a/LFM2-350M-Extract-TOON.Q5_K_M.gguf b/LFM2-350M-Extract-TOON.Q5_K_M.gguf new file mode 100644 index 0000000..e29f217 --- /dev/null +++ b/LFM2-350M-Extract-TOON.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e09e8a1fa149daeb55b7850b7f1cebeb1882247c7078e7eca1b357dc29db0213 +size 260375488 diff --git a/LFM2-350M-Extract-TOON.Q5_K_S.gguf b/LFM2-350M-Extract-TOON.Q5_K_S.gguf new file mode 100644 index 0000000..d3aacd4 --- /dev/null +++ b/LFM2-350M-Extract-TOON.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9374ebd68f56f853389d9cb64ed609e37abcd6dae32fd88a4061d88ef0c13c9 +size 255222720 diff --git a/LFM2-350M-Extract-TOON.Q6_K.gguf b/LFM2-350M-Extract-TOON.Q6_K.gguf new file mode 100644 index 0000000..f88b4d0 --- /dev/null +++ b/LFM2-350M-Extract-TOON.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4688f4351b47165fe4889319828086b4a41b937cbeeee28d9f69e5b1cfeb6481 +size 293381056 diff --git a/LFM2-350M-Extract-TOON.Q8_0.gguf b/LFM2-350M-Extract-TOON.Q8_0.gguf new file mode 100644 index 0000000..2164466 --- /dev/null +++ b/LFM2-350M-Extract-TOON.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39c7a4b141e914b05ac63bb2ab34090e955f81d4c765a027d237f7b4ed98f9dc +size 379216832 diff --git a/LFM2-350M-Extract-TOON.f16.gguf b/LFM2-350M-Extract-TOON.f16.gguf new file mode 100644 index 0000000..2c05ffc --- /dev/null +++ b/LFM2-350M-Extract-TOON.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4604fc79b733eb006bfed52fe540f4514ca0276009a584c1497a671de3b5b8f7 +size 711484352 diff --git a/README.md b/README.md new file mode 100644 index 0000000..173bbea --- /dev/null +++ b/README.md @@ -0,0 +1,83 @@ +--- +base_model: yasserrmd/LFM2-350M-Extract-TOON +datasets: +- yasserrmd/TOON-Unstructured-Structured +language: +- en +library_name: transformers +license: apache-2.0 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- text-generation +- instruction-tuning +- structured-output +- toon +- lfm2 +- unsloth +- lora +- transformers +--- +## About + + + + + + + + + +static quants of https://huggingface.co/yasserrmd/LFM2-350M-Extract-TOON + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#LFM2-350M-Extract-TOON-GGUF).*** + +weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.Q2_K.gguf) | Q2_K | 0.3 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.Q3_K_S.gguf) | Q3_K_S | 0.3 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.Q3_K_M.gguf) | Q3_K_M | 0.3 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.Q3_K_L.gguf) | Q3_K_L | 0.3 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.IQ4_XS.gguf) | IQ4_XS | 0.3 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.Q4_K_S.gguf) | Q4_K_S | 0.3 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.Q4_K_M.gguf) | Q4_K_M | 0.3 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.Q5_K_S.gguf) | Q5_K_S | 0.4 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.Q5_K_M.gguf) | Q5_K_M | 0.4 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.Q6_K.gguf) | Q6_K | 0.4 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.Q8_0.gguf) | Q8_0 | 0.5 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/LFM2-350M-Extract-TOON-GGUF/resolve/main/LFM2-350M-Extract-TOON.f16.gguf) | f16 | 0.8 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +