From fcc72e875b79230544c61a09c78b0970cf764f0d Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 21 Apr 2026 17:23:37 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: gaianet/Falcon3-1B-Instruct-GGUF Source: Original Platform --- .gitattributes | 48 +++++++++++++++++++++++++++++++++ Falcon3-1B-Instruct-Q2_K.gguf | 3 +++ Falcon3-1B-Instruct-Q3_K_L.gguf | 3 +++ Falcon3-1B-Instruct-Q3_K_M.gguf | 3 +++ Falcon3-1B-Instruct-Q3_K_S.gguf | 3 +++ Falcon3-1B-Instruct-Q4_0.gguf | 3 +++ Falcon3-1B-Instruct-Q4_K_M.gguf | 3 +++ Falcon3-1B-Instruct-Q4_K_S.gguf | 3 +++ Falcon3-1B-Instruct-Q5_0.gguf | 3 +++ Falcon3-1B-Instruct-Q5_K_M.gguf | 3 +++ Falcon3-1B-Instruct-Q5_K_S.gguf | 3 +++ Falcon3-1B-Instruct-Q6_K.gguf | 3 +++ Falcon3-1B-Instruct-Q8_0.gguf | 3 +++ Falcon3-1B-Instruct-f16.gguf | 3 +++ README.md | 41 ++++++++++++++++++++++++++++ config.json | 28 +++++++++++++++++++ 16 files changed, 156 insertions(+) create mode 100644 .gitattributes create mode 100644 Falcon3-1B-Instruct-Q2_K.gguf create mode 100644 Falcon3-1B-Instruct-Q3_K_L.gguf create mode 100644 Falcon3-1B-Instruct-Q3_K_M.gguf create mode 100644 Falcon3-1B-Instruct-Q3_K_S.gguf create mode 100644 Falcon3-1B-Instruct-Q4_0.gguf create mode 100644 Falcon3-1B-Instruct-Q4_K_M.gguf create mode 100644 Falcon3-1B-Instruct-Q4_K_S.gguf create mode 100644 Falcon3-1B-Instruct-Q5_0.gguf create mode 100644 Falcon3-1B-Instruct-Q5_K_M.gguf create mode 100644 Falcon3-1B-Instruct-Q5_K_S.gguf create mode 100644 Falcon3-1B-Instruct-Q6_K.gguf create mode 100644 Falcon3-1B-Instruct-Q8_0.gguf create mode 100644 Falcon3-1B-Instruct-f16.gguf create mode 100644 README.md create mode 100644 config.json diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..0aafc5e --- /dev/null +++ b/.gitattributes @@ -0,0 +1,48 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Falcon3-1B-Instruct-f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Falcon3-1B-Instruct-Q2_K.gguf b/Falcon3-1B-Instruct-Q2_K.gguf new file mode 100644 index 0000000..ff37569 --- /dev/null +++ b/Falcon3-1B-Instruct-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3045a784d1c15b8ff80d007c1a4113523c41dc0513e3d6fab6c909ced59dfb1c +size 727086976 diff --git a/Falcon3-1B-Instruct-Q3_K_L.gguf b/Falcon3-1B-Instruct-Q3_K_L.gguf new file mode 100644 index 0000000..ad0d546 --- /dev/null +++ b/Falcon3-1B-Instruct-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1e3efa32f2acb1679bdc2a177257c3634dd0db16a40b9005b2072e028dda844 +size 934246272 diff --git a/Falcon3-1B-Instruct-Q3_K_M.gguf b/Falcon3-1B-Instruct-Q3_K_M.gguf new file mode 100644 index 0000000..ed5ae55 --- /dev/null +++ b/Falcon3-1B-Instruct-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:af6a8a7261cbb82a219f8096a5071df5061fbb21289ce92219f045cc05e85947 +size 884963200 diff --git a/Falcon3-1B-Instruct-Q3_K_S.gguf b/Falcon3-1B-Instruct-Q3_K_S.gguf new file mode 100644 index 0000000..cfc44bb --- /dev/null +++ b/Falcon3-1B-Instruct-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e3c908860fb38d668b4ec02b05db9a292b2b1dad8a8c9879cc6c9bb0af1bbe0 +size 827193216 diff --git a/Falcon3-1B-Instruct-Q4_0.gguf b/Falcon3-1B-Instruct-Q4_0.gguf new file mode 100644 index 0000000..5076302 --- /dev/null +++ b/Falcon3-1B-Instruct-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec974b10a07d7614590d6e79a4f253ddfaa7b917c3556b4d4010912fd3d7f07b +size 1013249920 diff --git a/Falcon3-1B-Instruct-Q4_K_M.gguf b/Falcon3-1B-Instruct-Q4_K_M.gguf new file mode 100644 index 0000000..965a52a --- /dev/null +++ b/Falcon3-1B-Instruct-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:856e34792b706340545cd1b011d0d0a1a5a4b665da5f0de75d2a1e85eb4ac0d5 +size 1057044352 diff --git a/Falcon3-1B-Instruct-Q4_K_S.gguf b/Falcon3-1B-Instruct-Q4_K_S.gguf new file mode 100644 index 0000000..e2d3702 --- /dev/null +++ b/Falcon3-1B-Instruct-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3ed190b933870d7ce5ab568c528ea2195137aebd4fe67dd91d63b73cbcc36f2 +size 1018492800 diff --git a/Falcon3-1B-Instruct-Q5_0.gguf b/Falcon3-1B-Instruct-Q5_0.gguf new file mode 100644 index 0000000..afa15e2 --- /dev/null +++ b/Falcon3-1B-Instruct-Q5_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7cc700955186d92881b82aad1cc7a5bd0a2bd217bebe03d55aa95bfea81d414 +size 1188362112 diff --git a/Falcon3-1B-Instruct-Q5_K_M.gguf b/Falcon3-1B-Instruct-Q5_K_M.gguf new file mode 100644 index 0000000..a749211 --- /dev/null +++ b/Falcon3-1B-Instruct-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c70c9f17147416cd6ca26ca0f2f84b6f1278e03f3d742be71a915ad7b954c82 +size 1210922880 diff --git a/Falcon3-1B-Instruct-Q5_K_S.gguf b/Falcon3-1B-Instruct-Q5_K_S.gguf new file mode 100644 index 0000000..71c5e2b --- /dev/null +++ b/Falcon3-1B-Instruct-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a17d42e135cba6dfaa54e24a014bf6b1bac0990488c31b4515d5d5dfa11da6c4 +size 1188362112 diff --git a/Falcon3-1B-Instruct-Q6_K.gguf b/Falcon3-1B-Instruct-Q6_K.gguf new file mode 100644 index 0000000..160bbd6 --- /dev/null +++ b/Falcon3-1B-Instruct-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7d4a9f39614e3b74b78cedd48fc0595b09adb539e44154baaa379ff4f6bc578b +size 1374418816 diff --git a/Falcon3-1B-Instruct-Q8_0.gguf b/Falcon3-1B-Instruct-Q8_0.gguf new file mode 100644 index 0000000..0f57909 --- /dev/null +++ b/Falcon3-1B-Instruct-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd435ecb599339b89446df63ed78887d9b228a3dd3022a3c1f2c7bfdb9faa17d +size 1778710400 diff --git a/Falcon3-1B-Instruct-f16.gguf b/Falcon3-1B-Instruct-f16.gguf new file mode 100644 index 0000000..9dcdfaf --- /dev/null +++ b/Falcon3-1B-Instruct-f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2807c4440778122985679707d21759fe63d1212983800606cae8deb2cb449f0d +size 3343710080 diff --git a/README.md b/README.md new file mode 100644 index 0000000..81b1859 --- /dev/null +++ b/README.md @@ -0,0 +1,41 @@ +--- +base_model: tiiuae/Falcon3-1B-Instruct +license: other +license_name: falcon-llm-license +license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html +model_creator: tiiuae +model_name: Falcon3-1B-Instruct +quantized_by: Second State Inc. +library_name: transformers +language: +- en +- fr +- es +- pt +tags: +- falcon3 +--- + +# Falcon3-1B-Instruct-GGUF + +## Original Model + +[tiiuae/Falcon3-1B-Instruct](https://huggingface.co/tiiuae/Falcon3-1B-Instruct) + +## Run with Gaianet + +**Prompt template:** + +prompt template: `falcon3` + +**Context size:** + +chat_ctx_size: `8000` + +**Run with GaiaNet:** + +- Quick start: https://docs.gaianet.ai/node-guide/quick-start + +- Customize your node: https://docs.gaianet.ai/node-guide/customize + +*Quantized with llama.cpp b4381* \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000..fd9a3c8 --- /dev/null +++ b/config.json @@ -0,0 +1,28 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "eos_token_id": 11, + "head_dim": 256, + "hidden_act": "silu", + "hidden_size": 2048, + "initializer_range": 0.02, + "intermediate_size": 8192, + "max_position_embeddings": 8192, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 8, + "num_hidden_layers": 18, + "num_key_value_heads": 4, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_scaling": null, + "rope_theta": 1000042, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.46.1", + "use_cache": true, + "vocab_size": 131072 +}