commit d1ac30d4f7a918452cda5b5805808479a85f8945 Author: ModelHub XC Date: Wed May 6 01:40:41 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/LFM2-2.6B-mr-tictactoe-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..2b3dad4 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +LFM2-2.6B-mr-tictactoe.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/LFM2-2.6B-mr-tictactoe.IQ4_XS.gguf b/LFM2-2.6B-mr-tictactoe.IQ4_XS.gguf new file mode 100644 index 0000000..db4564d --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d591b99d5671c2b5797f649a98713094026d7bafe6a22503e1050fec6993466 +size 1417343968 diff --git a/LFM2-2.6B-mr-tictactoe.Q2_K.gguf b/LFM2-2.6B-mr-tictactoe.Q2_K.gguf new file mode 100644 index 0000000..43071ce --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8bb69a61e0049041ce7e4904c9b4f34ba81b1486f8c9b22b262a119f58870636 +size 983806944 diff --git a/LFM2-2.6B-mr-tictactoe.Q3_K_L.gguf b/LFM2-2.6B-mr-tictactoe.Q3_K_L.gguf new file mode 100644 index 0000000..6c44548 --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c54c764f0afe28a95ec379385a7b2e55f0ac76f4a69f5701d40a44aff2ad3121 +size 1340863456 diff --git a/LFM2-2.6B-mr-tictactoe.Q3_K_M.gguf b/LFM2-2.6B-mr-tictactoe.Q3_K_M.gguf new file mode 100644 index 0000000..2e144c6 --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f09cd45fd1d608aeb0604e77b0fbc373800ac93925f6865c919b2452c4387750 +size 1256059872 diff --git a/LFM2-2.6B-mr-tictactoe.Q3_K_S.gguf b/LFM2-2.6B-mr-tictactoe.Q3_K_S.gguf new file mode 100644 index 0000000..3aaeac1 --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d1a7de06e33903566419b1abfb88b5d1131e2e7c298af76ba0a614e14f24c7c +size 1159738336 diff --git a/LFM2-2.6B-mr-tictactoe.Q4_K_M.gguf b/LFM2-2.6B-mr-tictactoe.Q4_K_M.gguf new file mode 100644 index 0000000..8e5ab42 --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c2a6de31e3eee90722f819c1a9041841bedadff6ab7b9b99d2b4e88e18b203e +size 1563669472 diff --git a/LFM2-2.6B-mr-tictactoe.Q4_K_S.gguf b/LFM2-2.6B-mr-tictactoe.Q4_K_S.gguf new file mode 100644 index 0000000..ee09033 --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f13d649d46f50ba71f618ad91edc61dec0b0d081baa967e65bf146b20702551e +size 1491891168 diff --git a/LFM2-2.6B-mr-tictactoe.Q5_K_M.gguf b/LFM2-2.6B-mr-tictactoe.Q5_K_M.gguf new file mode 100644 index 0000000..4ef0c6b --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d77e9223e7f5f559b793d86618ad0a183b1b863d3fb109a97a4682c3f07fbd7b +size 1828959200 diff --git a/LFM2-2.6B-mr-tictactoe.Q5_K_S.gguf b/LFM2-2.6B-mr-tictactoe.Q5_K_S.gguf new file mode 100644 index 0000000..8735e2a --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e15de9489f3cb8ea9529ac59c630cdb5be8104f74c9edca64a944070a630b1a +size 1787458528 diff --git a/LFM2-2.6B-mr-tictactoe.Q6_K.gguf b/LFM2-2.6B-mr-tictactoe.Q6_K.gguf new file mode 100644 index 0000000..09d8021 --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29df91520287aa34ece6f0632201d5bf15b39b87aa66cafd37bf41bddf0b8586 +size 2110829536 diff --git a/LFM2-2.6B-mr-tictactoe.Q8_0.gguf b/LFM2-2.6B-mr-tictactoe.Q8_0.gguf new file mode 100644 index 0000000..1b2c4eb --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c16ed40213e20accaf1916681a7d717ee2c740bbcad818904b356735d062d6e +size 2733011936 diff --git a/LFM2-2.6B-mr-tictactoe.f16.gguf b/LFM2-2.6B-mr-tictactoe.f16.gguf new file mode 100644 index 0000000..38a5945 --- /dev/null +++ b/LFM2-2.6B-mr-tictactoe.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:186f3dcf4ee08c5c2a8a3fbc2e7b1ae009001ffff03871c04e391e3ffddd58c4 +size 5141459936 diff --git a/README.md b/README.md new file mode 100644 index 0000000..b055010 --- /dev/null +++ b/README.md @@ -0,0 +1,79 @@ +--- +base_model: anakin87/LFM2-2.6B-mr-tictactoe +language: +- en +library_name: transformers +license: other +license_link: https://www.liquid.ai/legal/lfm-license +license_name: lfm1.0 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- rl +- cispo +- tictactoe +- rlvr +--- +## About + + + + + + + + + +static quants of https://huggingface.co/anakin87/LFM2-2.6B-mr-tictactoe + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#LFM2-2.6B-mr-tictactoe-GGUF).*** + +weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.Q2_K.gguf) | Q2_K | 1.1 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.Q3_K_S.gguf) | Q3_K_S | 1.3 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.Q3_K_M.gguf) | Q3_K_M | 1.4 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.Q3_K_L.gguf) | Q3_K_L | 1.4 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.IQ4_XS.gguf) | IQ4_XS | 1.5 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.Q4_K_S.gguf) | Q4_K_S | 1.6 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.Q4_K_M.gguf) | Q4_K_M | 1.7 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.Q5_K_S.gguf) | Q5_K_S | 1.9 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.Q5_K_M.gguf) | Q5_K_M | 1.9 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.Q6_K.gguf) | Q6_K | 2.2 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.Q8_0.gguf) | Q8_0 | 2.8 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/LFM2-2.6B-mr-tictactoe-GGUF/resolve/main/LFM2-2.6B-mr-tictactoe.f16.gguf) | f16 | 5.2 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +