commit 0e20fffeef264839e666705a3731978684073d96 Author: ModelHub XC Date: Wed Apr 22 09:24:30 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/LFM2.5-1.2B-JP-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..6c4f0e5 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +LFM2.5-1.2B-JP.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/LFM2.5-1.2B-JP.IQ4_XS.gguf b/LFM2.5-1.2B-JP.IQ4_XS.gguf new file mode 100644 index 0000000..1cb9290 --- /dev/null +++ b/LFM2.5-1.2B-JP.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a0dc00728b79264f4c7a186a93fbd2fe38eccc1837ab975f2c0b11d6041ba07 +size 668620224 diff --git a/LFM2.5-1.2B-JP.Q2_K.gguf b/LFM2.5-1.2B-JP.Q2_K.gguf new file mode 100644 index 0000000..c52da04 --- /dev/null +++ b/LFM2.5-1.2B-JP.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4abc1ce1fad6f0c61c7f82425375a07af5b5028049430bf46ba4ebf0fb055ec +size 483399104 diff --git a/LFM2.5-1.2B-JP.Q3_K_L.gguf b/LFM2.5-1.2B-JP.Q3_K_L.gguf new file mode 100644 index 0000000..11b351a --- /dev/null +++ b/LFM2.5-1.2B-JP.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8df1c8c89cbd945a88c4f0919f756c05a62cd5c818e330d037122549db5e6503 +size 635475392 diff --git a/LFM2.5-1.2B-JP.Q3_K_M.gguf b/LFM2.5-1.2B-JP.Q3_K_M.gguf new file mode 100644 index 0000000..c3237ec --- /dev/null +++ b/LFM2.5-1.2B-JP.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14d0260083dbb4cfc416dbf58dbfa11b04c0139617b93d545d0352b72e99cfb3 +size 600348096 diff --git a/LFM2.5-1.2B-JP.Q3_K_S.gguf b/LFM2.5-1.2B-JP.Q3_K_S.gguf new file mode 100644 index 0000000..0def9d5 --- /dev/null +++ b/LFM2.5-1.2B-JP.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e99dac1c5d86c21a3c3d5b77871dc384e19bc39ad441ff96bb4818c36825947a +size 558159296 diff --git a/LFM2.5-1.2B-JP.Q4_K_M.gguf b/LFM2.5-1.2B-JP.Q4_K_M.gguf new file mode 100644 index 0000000..37b536f --- /dev/null +++ b/LFM2.5-1.2B-JP.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8061925f36c85b6e7c78414b7da10e1e98666d75b19b03afa8bc3ab1843cb728 +size 730895808 diff --git a/LFM2.5-1.2B-JP.Q4_K_S.gguf b/LFM2.5-1.2B-JP.Q4_K_S.gguf new file mode 100644 index 0000000..b508ada --- /dev/null +++ b/LFM2.5-1.2B-JP.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a36ef7cd4dd3d39014f20e9293976747c28ce6ac4392f949674f3468e4d23d8 +size 700470720 diff --git a/LFM2.5-1.2B-JP.Q5_K_M.gguf b/LFM2.5-1.2B-JP.Q5_K_M.gguf new file mode 100644 index 0000000..10cf42b --- /dev/null +++ b/LFM2.5-1.2B-JP.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a39df64fb293d046b58269ac62c45fdb6de172aa8305326bb7b8b12be575fca9 +size 843355584 diff --git a/LFM2.5-1.2B-JP.Q5_K_S.gguf b/LFM2.5-1.2B-JP.Q5_K_S.gguf new file mode 100644 index 0000000..8fcb99c --- /dev/null +++ b/LFM2.5-1.2B-JP.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd24240fcdb0de25edef32e2a43c950e2e4561bb98ce3c172bb61d31941eb07d +size 825251264 diff --git a/LFM2.5-1.2B-JP.Q6_K.gguf b/LFM2.5-1.2B-JP.Q6_K.gguf new file mode 100644 index 0000000..656d15b --- /dev/null +++ b/LFM2.5-1.2B-JP.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:040392eb93f058a0a1a891d814407684de78fa094a3931e3b74b3ad9e05caa7f +size 962844096 diff --git a/LFM2.5-1.2B-JP.Q8_0.gguf b/LFM2.5-1.2B-JP.Q8_0.gguf new file mode 100644 index 0000000..93174ff --- /dev/null +++ b/LFM2.5-1.2B-JP.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ed03a676b6b6587821794e969cb48bfb6a2262c47e2558e4712122b6f2742c4 +size 1246254528 diff --git a/LFM2.5-1.2B-JP.f16.gguf b/LFM2.5-1.2B-JP.f16.gguf new file mode 100644 index 0000000..a67661f --- /dev/null +++ b/LFM2.5-1.2B-JP.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:149991e8c3ea0fadd3f21440b566f9f096174d51f140f7438aaf5ab63c2e905e +size 2343327168 diff --git a/README.md b/README.md new file mode 100644 index 0000000..141377a --- /dev/null +++ b/README.md @@ -0,0 +1,79 @@ +--- +base_model: LiquidAI/LFM2.5-1.2B-JP +language: +- en +- ja +library_name: transformers +license: other +license_link: LICENSE +license_name: lfm1.0 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- liquid +- lfm2.5 +- edge +--- +## About + + + + + + + + + +static quants of https://huggingface.co/LiquidAI/LFM2.5-1.2B-JP + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#LFM2.5-1.2B-JP-GGUF).*** + +weighted/imatrix quants are available at https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-i1-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.Q2_K.gguf) | Q2_K | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.Q3_K_S.gguf) | Q3_K_S | 0.7 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.Q3_K_M.gguf) | Q3_K_M | 0.7 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.Q3_K_L.gguf) | Q3_K_L | 0.7 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.IQ4_XS.gguf) | IQ4_XS | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.Q4_K_S.gguf) | Q4_K_S | 0.8 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.Q4_K_M.gguf) | Q4_K_M | 0.8 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.Q5_K_S.gguf) | Q5_K_S | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.Q5_K_M.gguf) | Q5_K_M | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.Q6_K.gguf) | Q6_K | 1.1 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.Q8_0.gguf) | Q8_0 | 1.3 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/LFM2.5-1.2B-JP-GGUF/resolve/main/LFM2.5-1.2B-JP.f16.gguf) | f16 | 2.4 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +