初始化项目,由ModelHub XC社区提供模型
Model: mradermacher/Falcon3-3B-Base-i1-GGUF Source: Original Platform
This commit is contained in:
58
.gitattributes
vendored
Normal file
58
.gitattributes
vendored
Normal file
@@ -0,0 +1,58 @@
|
||||
*.7z filter=lfs diff=lfs merge=lfs -text
|
||||
*.arrow filter=lfs diff=lfs merge=lfs -text
|
||||
*.bin filter=lfs diff=lfs merge=lfs -text
|
||||
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
||||
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
||||
*.ftz filter=lfs diff=lfs merge=lfs -text
|
||||
*.gz filter=lfs diff=lfs merge=lfs -text
|
||||
*.h5 filter=lfs diff=lfs merge=lfs -text
|
||||
*.joblib filter=lfs diff=lfs merge=lfs -text
|
||||
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
||||
*.model filter=lfs diff=lfs merge=lfs -text
|
||||
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
||||
*.npy filter=lfs diff=lfs merge=lfs -text
|
||||
*.npz filter=lfs diff=lfs merge=lfs -text
|
||||
*.onnx filter=lfs diff=lfs merge=lfs -text
|
||||
*.ot filter=lfs diff=lfs merge=lfs -text
|
||||
*.parquet filter=lfs diff=lfs merge=lfs -text
|
||||
*.pb filter=lfs diff=lfs merge=lfs -text
|
||||
*.pickle filter=lfs diff=lfs merge=lfs -text
|
||||
*.pkl filter=lfs diff=lfs merge=lfs -text
|
||||
*.pt filter=lfs diff=lfs merge=lfs -text
|
||||
*.pth filter=lfs diff=lfs merge=lfs -text
|
||||
*.rar filter=lfs diff=lfs merge=lfs -text
|
||||
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
||||
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar filter=lfs diff=lfs merge=lfs -text
|
||||
*.tflite filter=lfs diff=lfs merge=lfs -text
|
||||
*.tgz filter=lfs diff=lfs merge=lfs -text
|
||||
*.wasm filter=lfs diff=lfs merge=lfs -text
|
||||
*.xz filter=lfs diff=lfs merge=lfs -text
|
||||
*.zip filter=lfs diff=lfs merge=lfs -text
|
||||
*.zst filter=lfs diff=lfs merge=lfs -text
|
||||
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||
imatrix.dat filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Falcon3-3B-Base.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
3
Falcon3-3B-Base.i1-IQ1_M.gguf
Normal file
3
Falcon3-3B-Base.i1-IQ1_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:677e05b38a94d75876f957089a8cfac025a4d87ccbd6029bd4fba2c91edeb2fa
|
||||
size 965970464
|
||||
3
Falcon3-3B-Base.i1-IQ1_S.gguf
Normal file
3
Falcon3-3B-Base.i1-IQ1_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:4626e54ca47df35d962a2668e404fc48c0cd6b29e6ed35f6a2b300127f1b6699
|
||||
size 917015072
|
||||
3
Falcon3-3B-Base.i1-IQ2_M.gguf
Normal file
3
Falcon3-3B-Base.i1-IQ2_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:8ce3ebb4042b3d6740b10728c0ced248e03ff763ceb852f6fd40558eca51944c
|
||||
size 1267468832
|
||||
3
Falcon3-3B-Base.i1-IQ2_S.gguf
Normal file
3
Falcon3-3B-Base.i1-IQ2_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:ffca6398b85957f16078c95ec7f82dc1fcf402717f508bc434e3baa0986294a3
|
||||
size 1202194976
|
||||
3
Falcon3-3B-Base.i1-IQ2_XS.gguf
Normal file
3
Falcon3-3B-Base.i1-IQ2_XS.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:f757fd1bb54f68ee2474f7ca21944dc45bb68c413bca2a09d60550c4dc3a0463
|
||||
size 1119324704
|
||||
3
Falcon3-3B-Base.i1-IQ2_XXS.gguf
Normal file
3
Falcon3-3B-Base.i1-IQ2_XXS.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:cac0a9bfb7967e9639827984ec13157441676a3a7d86c46177e3240d081abe0d
|
||||
size 1047562784
|
||||
3
Falcon3-3B-Base.i1-IQ3_M.gguf
Normal file
3
Falcon3-3B-Base.i1-IQ3_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:578b992f4796ca989d2afe39bb9bb48b511f3a977b17d27bc19e42d128bea60c
|
||||
size 1593690656
|
||||
3
Falcon3-3B-Base.i1-IQ3_S.gguf
Normal file
3
Falcon3-3B-Base.i1-IQ3_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:e924fbab8157f9d89a0a2e501b3ccfce6f8fd9e64a12a38b6cada39ebbc30607
|
||||
size 1549404704
|
||||
3
Falcon3-3B-Base.i1-IQ3_XS.gguf
Normal file
3
Falcon3-3B-Base.i1-IQ3_XS.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:877b5e18ec295de030597e71e20165e3c1dd41c268073742a599ed4f8614d760
|
||||
size 1491307040
|
||||
3
Falcon3-3B-Base.i1-IQ3_XXS.gguf
Normal file
3
Falcon3-3B-Base.i1-IQ3_XXS.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:2346bb904609906e02baad90121768a37a40b877d06e876002d47dc251f7e79b
|
||||
size 1374816800
|
||||
3
Falcon3-3B-Base.i1-IQ4_XS.gguf
Normal file
3
Falcon3-3B-Base.i1-IQ4_XS.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:6621947782504f1b0b482fdad6ea98d65d559bd46f5b2e9204e3b0afdcfda259
|
||||
size 1836304928
|
||||
3
Falcon3-3B-Base.i1-Q2_K.gguf
Normal file
3
Falcon3-3B-Base.i1-Q2_K.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:6c7412d321ad9f423e1ea4e47ed7bd741e67152b54d1f00015ae65cd4014863f
|
||||
size 1353878048
|
||||
3
Falcon3-3B-Base.i1-Q2_K_S.gguf
Normal file
3
Falcon3-3B-Base.i1-Q2_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:2e47ddd9fda938e41bcc2459f5b2dcd3243f25800d2027c8738c2390a0100993
|
||||
size 1275775520
|
||||
3
Falcon3-3B-Base.i1-Q3_K_L.gguf
Normal file
3
Falcon3-3B-Base.i1-Q3_K_L.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:12a607632843a71bcd394b613778e95b3baed0b331529f6b96446b45e6974fc8
|
||||
size 1781352992
|
||||
3
Falcon3-3B-Base.i1-Q3_K_M.gguf
Normal file
3
Falcon3-3B-Base.i1-Q3_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:f8dcf4e0cd2a20f250ca8303e70a214e1f8678cbf89824a868075f16e87f5afc
|
||||
size 1673218592
|
||||
3
Falcon3-3B-Base.i1-Q3_K_S.gguf
Normal file
3
Falcon3-3B-Base.i1-Q3_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:80a11b9ffb8a011645832733316e4dc9e7d69291bd5235f63d0a72117b66a9c0
|
||||
size 1549404704
|
||||
3
Falcon3-3B-Base.i1-Q4_0.gguf
Normal file
3
Falcon3-3B-Base.i1-Q4_0.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:6d4d0976604a42795adb274b0ac9ff551e6e346d89b30cc3d5d3d5a01a26c9a6
|
||||
size 1928120864
|
||||
3
Falcon3-3B-Base.i1-Q4_K_M.gguf
Normal file
3
Falcon3-3B-Base.i1-Q4_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:c2e80e4b76ad47506beb40a982cd75e69f2fb4e17070349970afe8bc9ac28e9c
|
||||
size 2005682720
|
||||
3
Falcon3-3B-Base.i1-Q4_K_S.gguf
Normal file
3
Falcon3-3B-Base.i1-Q4_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:2cdd84df21d3e67ae081f66fe622c2a63521bd01088bee2ce8c5307e065463b6
|
||||
size 1933232672
|
||||
3
Falcon3-3B-Base.i1-Q5_K_M.gguf
Normal file
3
Falcon3-3B-Base.i1-Q5_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:351965b239c5d80aa40e162a5efba5b33bb75b19d512225be045834ef446d692
|
||||
size 2319469088
|
||||
3
Falcon3-3B-Base.i1-Q5_K_S.gguf
Normal file
3
Falcon3-3B-Base.i1-Q5_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:cf461dcff9cf51ff788843b57e3dbc0f4dbf02bd5442e0adb8f927076b5d098f
|
||||
size 2277689888
|
||||
3
Falcon3-3B-Base.i1-Q6_K.gguf
Normal file
3
Falcon3-3B-Base.i1-Q6_K.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:d70d015e4e3d156db58b2c0e381381db6b61534b34510c63c9b36e484b9b9e56
|
||||
size 2652867104
|
||||
81
README.md
Normal file
81
README.md
Normal file
@@ -0,0 +1,81 @@
|
||||
---
|
||||
base_model: tiiuae/Falcon3-3B-Base
|
||||
language:
|
||||
- en
|
||||
- fr
|
||||
- es
|
||||
- pt
|
||||
library_name: transformers
|
||||
license: other
|
||||
license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
|
||||
license_name: falcon-llm-license
|
||||
quantized_by: mradermacher
|
||||
tags:
|
||||
- falcon3
|
||||
---
|
||||
## About
|
||||
|
||||
<!-- ### quantize_version: 2 -->
|
||||
<!-- ### output_tensor_quantised: 1 -->
|
||||
<!-- ### convert_type: hf -->
|
||||
<!-- ### vocab_type: -->
|
||||
<!-- ### tags: nicoboss -->
|
||||
weighted/imatrix quants of https://huggingface.co/tiiuae/Falcon3-3B-Base
|
||||
|
||||
<!-- provided-files -->
|
||||
static quants are available at https://huggingface.co/mradermacher/Falcon3-3B-Base-GGUF
|
||||
## Usage
|
||||
|
||||
If you are unsure how to use GGUF files, refer to one of [TheBloke's
|
||||
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
|
||||
more details, including on how to concatenate multi-part files.
|
||||
|
||||
## Provided Quants
|
||||
|
||||
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
|
||||
|
||||
| Link | Type | Size/GB | Notes |
|
||||
|:-----|:-----|--------:|:------|
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ1_S.gguf) | i1-IQ1_S | 1.0 | for the desperate |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ1_M.gguf) | i1-IQ1_M | 1.1 | mostly desperate |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 1.1 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ2_XS.gguf) | i1-IQ2_XS | 1.2 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ2_S.gguf) | i1-IQ2_S | 1.3 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ2_M.gguf) | i1-IQ2_M | 1.4 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q2_K_S.gguf) | i1-Q2_K_S | 1.4 | very low quality |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q2_K.gguf) | i1-Q2_K | 1.5 | IQ3_XXS probably better |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 1.5 | lower quality |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ3_XS.gguf) | i1-IQ3_XS | 1.6 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ3_S.gguf) | i1-IQ3_S | 1.6 | beats Q3_K* |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q3_K_S.gguf) | i1-Q3_K_S | 1.6 | IQ3_XS probably better |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ3_M.gguf) | i1-IQ3_M | 1.7 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q3_K_M.gguf) | i1-Q3_K_M | 1.8 | IQ3_S probably better |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q3_K_L.gguf) | i1-Q3_K_L | 1.9 | IQ3_M probably better |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-IQ4_XS.gguf) | i1-IQ4_XS | 1.9 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q4_0.gguf) | i1-Q4_0 | 2.0 | fast, low quality |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q4_K_S.gguf) | i1-Q4_K_S | 2.0 | optimal size/speed/quality |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q4_K_M.gguf) | i1-Q4_K_M | 2.1 | fast, recommended |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q5_K_S.gguf) | i1-Q5_K_S | 2.4 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q5_K_M.gguf) | i1-Q5_K_M | 2.4 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/Falcon3-3B-Base-i1-GGUF/resolve/main/Falcon3-3B-Base.i1-Q6_K.gguf) | i1-Q6_K | 2.8 | practically like static Q6_K |
|
||||
|
||||
Here is a handy graph by ikawrakow comparing some lower-quality quant
|
||||
types (lower is better):
|
||||
|
||||

|
||||
|
||||
And here are Artefact2's thoughts on the matter:
|
||||
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
|
||||
|
||||
## FAQ / Model Request
|
||||
|
||||
See https://huggingface.co/mradermacher/model_requests for some answers to
|
||||
questions you might have and/or if you want some other model quantized.
|
||||
|
||||
## Thanks
|
||||
|
||||
I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
|
||||
me use its servers and providing upgrades to my workstation to enable
|
||||
this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.
|
||||
|
||||
<!-- end -->
|
||||
3
imatrix.dat
Normal file
3
imatrix.dat
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:d913ee28a4616b899baa24fad0243baf75b7f8df9df9dd8055b3ab2116589a57
|
||||
size 2438115
|
||||
Reference in New Issue
Block a user