commit 0be05e52f0d3f0615ba4ae90054d4ecdf26bc447 Author: ModelHub XC Date: Fri Apr 10 10:59:12 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/Falcon3-1B-Base-i1-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..8fe17c6 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,58 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +imatrix.dat filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Base.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Falcon3-1B-Base.i1-IQ1_M.gguf b/Falcon3-1B-Base.i1-IQ1_M.gguf new file mode 100644 index 0000000..e77672e --- /dev/null +++ b/Falcon3-1B-Base.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:318609af005da4b9b09c3b596b52019b047a0d07be4f5ad8b03e67b6ced1de58 +size 536146368 diff --git a/Falcon3-1B-Base.i1-IQ1_S.gguf b/Falcon3-1B-Base.i1-IQ1_S.gguf new file mode 100644 index 0000000..7022986 --- /dev/null +++ b/Falcon3-1B-Base.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad7617d0c94f98cb9d04de8c0cc43275a15c9d3860abf44df01a8e12dea48c15 +size 513044928 diff --git a/Falcon3-1B-Base.i1-IQ2_M.gguf b/Falcon3-1B-Base.i1-IQ2_M.gguf new file mode 100644 index 0000000..beb7d60 --- /dev/null +++ b/Falcon3-1B-Base.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7fb81322524a6e878a7bdacc0ddb20090d70b1b9bc5b5331c0748f8164550ac +size 683733440 diff --git a/Falcon3-1B-Base.i1-IQ2_S.gguf b/Falcon3-1B-Base.i1-IQ2_S.gguf new file mode 100644 index 0000000..f7cc2de --- /dev/null +++ b/Falcon3-1B-Base.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23d9eb45004078d4ab1a311b38e83cf7d718ae396dda041d9ede89e5ca17df37 +size 652931520 diff --git a/Falcon3-1B-Base.i1-IQ2_XS.gguf b/Falcon3-1B-Base.i1-IQ2_XS.gguf new file mode 100644 index 0000000..8f62136 --- /dev/null +++ b/Falcon3-1B-Base.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89e447db6a489d138e089d5f4b73c756b5a0b0a10930749d171130ed2ff191c8 +size 607809984 diff --git a/Falcon3-1B-Base.i1-IQ2_XXS.gguf b/Falcon3-1B-Base.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..0f2c3eb --- /dev/null +++ b/Falcon3-1B-Base.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10b2d03cf1942c422fe913c7551796f6a3a3ae31c15d468efdde080de38e846b +size 574648768 diff --git a/Falcon3-1B-Base.i1-IQ3_M.gguf b/Falcon3-1B-Base.i1-IQ3_M.gguf new file mode 100644 index 0000000..790fbce --- /dev/null +++ b/Falcon3-1B-Base.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcf64d7d3605aa96965e722406ac758a04f2fe266a52218dc4fe358c78d18b4b +size 846688704 diff --git a/Falcon3-1B-Base.i1-IQ3_S.gguf b/Falcon3-1B-Base.i1-IQ3_S.gguf new file mode 100644 index 0000000..2a511e3 --- /dev/null +++ b/Falcon3-1B-Base.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa5a31d2dbb7bf56736b9e93ac34b10e5bfbdedc6ff052d1004b5e948368db8f +size 827191744 diff --git a/Falcon3-1B-Base.i1-IQ3_XS.gguf b/Falcon3-1B-Base.i1-IQ3_XS.gguf new file mode 100644 index 0000000..3fb9656 --- /dev/null +++ b/Falcon3-1B-Base.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:083270e80c19f9bc9c65334accb13a2b9ef43de2e99bc7e723a4e75f87f8f11e +size 801436096 diff --git a/Falcon3-1B-Base.i1-IQ3_XXS.gguf b/Falcon3-1B-Base.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..613280e --- /dev/null +++ b/Falcon3-1B-Base.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f31de3460d05cdaa1a81c84161e2c820d732474bd005ee580af2d53ad3d7149 +size 734917056 diff --git a/Falcon3-1B-Base.i1-IQ4_XS.gguf b/Falcon3-1B-Base.i1-IQ4_XS.gguf new file mode 100644 index 0000000..12ece3c --- /dev/null +++ b/Falcon3-1B-Base.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ceceed0db544303f94dfda20ee37ee0f13ffe09431286ff11ccd7a39dcb7b81 +size 969470400 diff --git a/Falcon3-1B-Base.i1-Q2_K.gguf b/Falcon3-1B-Base.i1-Q2_K.gguf new file mode 100644 index 0000000..49036bb --- /dev/null +++ b/Falcon3-1B-Base.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe87201b66d4a89946372b2c346a1bd114254d19669274d6d8713a0371b39b93 +size 727085504 diff --git a/Falcon3-1B-Base.i1-Q2_K_S.gguf b/Falcon3-1B-Base.i1-Q2_K_S.gguf new file mode 100644 index 0000000..03870be --- /dev/null +++ b/Falcon3-1B-Base.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce628a088ae37d1118dfde66a178021fcd3faf475b51f5cd6d69e2f5cde3951c +size 692777408 diff --git a/Falcon3-1B-Base.i1-Q3_K_L.gguf b/Falcon3-1B-Base.i1-Q3_K_L.gguf new file mode 100644 index 0000000..5677b49 --- /dev/null +++ b/Falcon3-1B-Base.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8b734970b413c745f0fa4cea66840f2de6856b7bcf75f96dbc0ff6fa860cf0d +size 934244800 diff --git a/Falcon3-1B-Base.i1-Q3_K_M.gguf b/Falcon3-1B-Base.i1-Q3_K_M.gguf new file mode 100644 index 0000000..2e8d5a5 --- /dev/null +++ b/Falcon3-1B-Base.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0fe4343fcb04a09134d7ff0363a52736b7fa12a64e36b4a245c8eb6a7c7acaa2 +size 884961728 diff --git a/Falcon3-1B-Base.i1-Q3_K_S.gguf b/Falcon3-1B-Base.i1-Q3_K_S.gguf new file mode 100644 index 0000000..8fedf1e --- /dev/null +++ b/Falcon3-1B-Base.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94412d2346f3da212dc6aca42abd2800de75e593136a70395a534fd6db4bfdaf +size 827191744 diff --git a/Falcon3-1B-Base.i1-Q4_0.gguf b/Falcon3-1B-Base.i1-Q4_0.gguf new file mode 100644 index 0000000..ac35412 --- /dev/null +++ b/Falcon3-1B-Base.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab47c3d0cc52531ffd98982cbcfba75de65aaf69d960a97d6401f80a1b5455e2 +size 1015345600 diff --git a/Falcon3-1B-Base.i1-Q4_K_M.gguf b/Falcon3-1B-Base.i1-Q4_K_M.gguf new file mode 100644 index 0000000..420bef3 --- /dev/null +++ b/Falcon3-1B-Base.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fa080be2dbdd3e8e78b66ab04f66421e70726139a38ee3dba3afe3c3c36c953 +size 1057042880 diff --git a/Falcon3-1B-Base.i1-Q4_K_S.gguf b/Falcon3-1B-Base.i1-Q4_K_S.gguf new file mode 100644 index 0000000..6f4857b --- /dev/null +++ b/Falcon3-1B-Base.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89a805a172d52d478dfa14bcab031c8c030e64d2ccc9ffbec12cf5e28bce2f0a +size 1018491328 diff --git a/Falcon3-1B-Base.i1-Q5_K_M.gguf b/Falcon3-1B-Base.i1-Q5_K_M.gguf new file mode 100644 index 0000000..1a23f0d --- /dev/null +++ b/Falcon3-1B-Base.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b24f2119aa2454be858e1d256949e6d10df2de034bfd12921caea65e736953a0 +size 1210921408 diff --git a/Falcon3-1B-Base.i1-Q5_K_S.gguf b/Falcon3-1B-Base.i1-Q5_K_S.gguf new file mode 100644 index 0000000..2671920 --- /dev/null +++ b/Falcon3-1B-Base.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:075d822aea52dd6cbfad100eea9498403fa26fb134bd54a765191f9338c04809 +size 1188360640 diff --git a/Falcon3-1B-Base.i1-Q6_K.gguf b/Falcon3-1B-Base.i1-Q6_K.gguf new file mode 100644 index 0000000..213004b --- /dev/null +++ b/Falcon3-1B-Base.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0fc9cb27c54cb018e2fe7e59561ca20629acc7572de9f343ee496063662ef6e2 +size 1374417344 diff --git a/README.md b/README.md new file mode 100644 index 0000000..eea972e --- /dev/null +++ b/README.md @@ -0,0 +1,81 @@ +--- +base_model: tiiuae/Falcon3-1B-Base +language: +- en +- fr +- es +- pt +library_name: transformers +license: other +license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html +license_name: falcon-llm-license +quantized_by: mradermacher +tags: +- falcon3 +--- +## About + + + + + + +weighted/imatrix quants of https://huggingface.co/tiiuae/Falcon3-1B-Base + + +static quants are available at https://huggingface.co/mradermacher/Falcon3-1B-Base-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-IQ1_S.gguf) | i1-IQ1_S | 0.6 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-IQ1_M.gguf) | i1-IQ1_M | 0.6 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 0.7 | | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-IQ2_XS.gguf) | i1-IQ2_XS | 0.7 | | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-IQ2_S.gguf) | i1-IQ2_S | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-IQ2_M.gguf) | i1-IQ2_M | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-Q2_K_S.gguf) | i1-Q2_K_S | 0.8 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-Q2_K.gguf) | i1-Q2_K | 0.8 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 0.8 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-IQ3_XS.gguf) | i1-IQ3_XS | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-IQ3_S.gguf) | i1-IQ3_S | 0.9 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-Q3_K_S.gguf) | i1-Q3_K_S | 0.9 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-IQ3_M.gguf) | i1-IQ3_M | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-Q3_K_M.gguf) | i1-Q3_K_M | 1.0 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-Q3_K_L.gguf) | i1-Q3_K_L | 1.0 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-IQ4_XS.gguf) | i1-IQ4_XS | 1.1 | | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-Q4_0.gguf) | i1-Q4_0 | 1.1 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-Q4_K_S.gguf) | i1-Q4_K_S | 1.1 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-Q4_K_M.gguf) | i1-Q4_K_M | 1.2 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-Q5_K_S.gguf) | i1-Q5_K_S | 1.3 | | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-Q5_K_M.gguf) | i1-Q5_K_M | 1.3 | | +| [GGUF](https://huggingface.co/mradermacher/Falcon3-1B-Base-i1-GGUF/resolve/main/Falcon3-1B-Base.i1-Q6_K.gguf) | i1-Q6_K | 1.5 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/imatrix.dat b/imatrix.dat new file mode 100644 index 0000000..bb40cae --- /dev/null +++ b/imatrix.dat @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47302623286f16ef6ee05587fcf9202ba2c8bb55c479709e5389676bfb9a96a8 +size 1478719