初始化项目,由ModelHub XC社区提供模型

Model: mradermacher/Falcon3-3B-Base-i1-GGUF
Source: Original Platform
This commit is contained in:
ModelHub XC
2026-05-03 10:33:44 +08:00
commit e848df6726
25 changed files with 208 additions and 0 deletions

58
.gitattributes vendored Normal file
View File

@@ -0,0 +1,58 @@
*.7z filter=lfs diff=lfs merge=lfs -text
*.arrow filter=lfs diff=lfs merge=lfs -text
*.bin filter=lfs diff=lfs merge=lfs -text
*.bz2 filter=lfs diff=lfs merge=lfs -text
*.ckpt filter=lfs diff=lfs merge=lfs -text
*.ftz filter=lfs diff=lfs merge=lfs -text
*.gz filter=lfs diff=lfs merge=lfs -text
*.h5 filter=lfs diff=lfs merge=lfs -text
*.joblib filter=lfs diff=lfs merge=lfs -text
*.lfs.* filter=lfs diff=lfs merge=lfs -text
*.mlmodel filter=lfs diff=lfs merge=lfs -text
*.model filter=lfs diff=lfs merge=lfs -text
*.msgpack filter=lfs diff=lfs merge=lfs -text
*.npy filter=lfs diff=lfs merge=lfs -text
*.npz filter=lfs diff=lfs merge=lfs -text
*.onnx filter=lfs diff=lfs merge=lfs -text
*.ot filter=lfs diff=lfs merge=lfs -text
*.parquet filter=lfs diff=lfs merge=lfs -text
*.pb filter=lfs diff=lfs merge=lfs -text
*.pickle filter=lfs diff=lfs merge=lfs -text
*.pkl filter=lfs diff=lfs merge=lfs -text
*.pt filter=lfs diff=lfs merge=lfs -text
*.pth filter=lfs diff=lfs merge=lfs -text
*.rar filter=lfs diff=lfs merge=lfs -text
*.safetensors filter=lfs diff=lfs merge=lfs -text
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
*.tar.* filter=lfs diff=lfs merge=lfs -text
*.tar filter=lfs diff=lfs merge=lfs -text
*.tflite filter=lfs diff=lfs merge=lfs -text
*.tgz filter=lfs diff=lfs merge=lfs -text
*.wasm filter=lfs diff=lfs merge=lfs -text
*.xz filter=lfs diff=lfs merge=lfs -text
*.zip filter=lfs diff=lfs merge=lfs -text
*.zst filter=lfs diff=lfs merge=lfs -text
*tfevents* filter=lfs diff=lfs merge=lfs -text
imatrix.dat filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
Falcon3-3B-Base.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:677e05b38a94d75876f957089a8cfac025a4d87ccbd6029bd4fba2c91edeb2fa
size 965970464

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:4626e54ca47df35d962a2668e404fc48c0cd6b29e6ed35f6a2b300127f1b6699
size 917015072

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:8ce3ebb4042b3d6740b10728c0ced248e03ff763ceb852f6fd40558eca51944c
size 1267468832

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:ffca6398b85957f16078c95ec7f82dc1fcf402717f508bc434e3baa0986294a3
size 1202194976

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:f757fd1bb54f68ee2474f7ca21944dc45bb68c413bca2a09d60550c4dc3a0463
size 1119324704

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:cac0a9bfb7967e9639827984ec13157441676a3a7d86c46177e3240d081abe0d
size 1047562784

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:578b992f4796ca989d2afe39bb9bb48b511f3a977b17d27bc19e42d128bea60c
size 1593690656

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:e924fbab8157f9d89a0a2e501b3ccfce6f8fd9e64a12a38b6cada39ebbc30607
size 1549404704

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:877b5e18ec295de030597e71e20165e3c1dd41c268073742a599ed4f8614d760
size 1491307040

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:2346bb904609906e02baad90121768a37a40b877d06e876002d47dc251f7e79b
size 1374816800

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:6621947782504f1b0b482fdad6ea98d65d559bd46f5b2e9204e3b0afdcfda259
size 1836304928

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:6c7412d321ad9f423e1ea4e47ed7bd741e67152b54d1f00015ae65cd4014863f
size 1353878048

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:2e47ddd9fda938e41bcc2459f5b2dcd3243f25800d2027c8738c2390a0100993
size 1275775520

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:12a607632843a71bcd394b613778e95b3baed0b331529f6b96446b45e6974fc8
size 1781352992

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:f8dcf4e0cd2a20f250ca8303e70a214e1f8678cbf89824a868075f16e87f5afc
size 1673218592

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:80a11b9ffb8a011645832733316e4dc9e7d69291bd5235f63d0a72117b66a9c0
size 1549404704

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:6d4d0976604a42795adb274b0ac9ff551e6e346d89b30cc3d5d3d5a01a26c9a6
size 1928120864

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:c2e80e4b76ad47506beb40a982cd75e69f2fb4e17070349970afe8bc9ac28e9c
size 2005682720

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:2cdd84df21d3e67ae081f66fe622c2a63521bd01088bee2ce8c5307e065463b6
size 1933232672

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:351965b239c5d80aa40e162a5efba5b33bb75b19d512225be045834ef446d692
size 2319469088

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:cf461dcff9cf51ff788843b57e3dbc0f4dbf02bd5442e0adb8f927076b5d098f
size 2277689888

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:d70d015e4e3d156db58b2c0e381381db6b61534b34510c63c9b36e484b9b9e56
size 2652867104

81
README.md Normal file
View File

@@ -0,0 +1,81 @@
---
base_model: tiiuae/Falcon3-3B-Base
language:
- en
- fr
- es
- pt
library_name: transformers
license: other
license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
license_name: falcon-llm-license
quantized_by: mradermacher
tags:
- falcon3
---
## About
<!-- ### quantize_version: 2 -->
<!-- ### output_tensor_quantised: 1 -->
<!-- ### convert_type: hf -->
<!-- ### vocab_type: -->
<!-- ### tags: nicoboss -->
weighted/imatrix quants of https://huggingface.co/tiiuae/Falcon3-3B-Base
<!-- provided-files -->
static quants are available at https://huggingface.co/mradermacher/Falcon3-3B-Base-GGUF
## Usage
If you are unsure how to use GGUF files, refer to one of [TheBloke's
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
more details, including on how to concatenate multi-part files.
## Provided Quants
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
| Link | Type | Size/GB | Notes |
|:-----|:-----|--------:|:------|
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ1_S.gguf) | i1-IQ1_S | 1.0 | for the desperate |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ1_M.gguf) | i1-IQ1_M | 1.1 | mostly desperate |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 1.1 | |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ2_XS.gguf) | i1-IQ2_XS | 1.2 | |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ2_S.gguf) | i1-IQ2_S | 1.3 | |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ2_M.gguf) | i1-IQ2_M | 1.4 | |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q2_K_S.gguf) | i1-Q2_K_S | 1.4 | very low quality |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q2_K.gguf) | i1-Q2_K | 1.5 | IQ3_XXS probably better |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 1.5 | lower quality |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ3_XS.gguf) | i1-IQ3_XS | 1.6 | |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ3_S.gguf) | i1-IQ3_S | 1.6 | beats Q3_K* |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q3_K_S.gguf) | i1-Q3_K_S | 1.6 | IQ3_XS probably better |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ3_M.gguf) | i1-IQ3_M | 1.7 | |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q3_K_M.gguf) | i1-Q3_K_M | 1.8 | IQ3_S probably better |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q3_K_L.gguf) | i1-Q3_K_L | 1.9 | IQ3_M probably better |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ4_XS.gguf) | i1-IQ4_XS | 1.9 | |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q4_0.gguf) | i1-Q4_0 | 2.0 | fast, low quality |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q4_K_S.gguf) | i1-Q4_K_S | 2.0 | optimal size/speed/quality |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q4_K_M.gguf) | i1-Q4_K_M | 2.1 | fast, recommended |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q5_K_S.gguf) | i1-Q5_K_S | 2.4 | |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q5_K_M.gguf) | i1-Q5_K_M | 2.4 | |
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q6_K.gguf) | i1-Q6_K | 2.8 | practically like static Q6_K |
Here is a handy graph by ikawrakow comparing some lower-quality quant
types (lower is better):
![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)
And here are Artefact2's thoughts on the matter:
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
## FAQ / Model Request
See https://huggingface.co/mradermacher/model_requests for some answers to
questions you might have and/or if you want some other model quantized.
## Thanks
I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
me use its servers and providing upgrades to my workstation to enable
this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.
<!-- end -->

3
imatrix.dat Normal file
View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:d913ee28a4616b899baa24fad0243baf75b7f8df9df9dd8055b3ab2116589a57
size 2438115