Update metadata with huggingface_hub

This commit is contained in:
ai-modelscope
2024-12-23 19:06:02 +08:00
parent 337dc56af0
commit 4143f44cf4
25 changed files with 147 additions and 52 deletions

106
README.md
View File

@@ -1,17 +1,112 @@
---
quantized_by: bartowski
pipeline_tag: text-generation
license_name: falcon-llm-license
tags:
- falcon3
license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
license: other
base_model: tiiuae/Falcon3-10B-Instruct
license_name: falcon-llm-license
license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
model-index:
- name: Falcon3-10B-Instruct
results:
- task:
type: text-generation
name: Text Generation
dataset:
name: IFEval (0-Shot)
type: HuggingFaceH4/ifeval
args:
num_few_shot: 0
metrics:
- type: inst_level_strict_acc and prompt_level_strict_acc
value: 78.17
name: strict accuracy
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=tiiuae/Falcon3-10B-Instruct
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: BBH (3-Shot)
type: BBH
args:
num_few_shot: 3
metrics:
- type: acc_norm
value: 44.82
name: normalized accuracy
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=tiiuae/Falcon3-10B-Instruct
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MATH Lvl 5 (4-Shot)
type: hendrycks/competition_math
args:
num_few_shot: 4
metrics:
- type: exact_match
value: 25.91
name: exact match
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=tiiuae/Falcon3-10B-Instruct
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: GPQA (0-shot)
type: Idavidrein/gpqa
args:
num_few_shot: 0
metrics:
- type: acc_norm
value: 10.51
name: acc_norm
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=tiiuae/Falcon3-10B-Instruct
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MuSR (0-shot)
type: TAUR-Lab/MuSR
args:
num_few_shot: 0
metrics:
- type: acc_norm
value: 13.61
name: acc_norm
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=tiiuae/Falcon3-10B-Instruct
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MMLU-PRO (5-shot)
type: TIGER-Lab/MMLU-Pro
config: main
split: test
args:
num_few_shot: 5
metrics:
- type: acc
value: 38.1
name: accuracy
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=tiiuae/Falcon3-10B-Instruct
name: Open LLM Leaderboard
---
## Llamacpp imatrix Quantizations of Falcon3-10B-Instruct
Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b4341">b4341</a> for quantization.
Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b4381">b4381</a> for quantization.
Original model: https://huggingface.co/tiiuae/Falcon3-10B-Instruct
@@ -29,11 +124,14 @@ Run them in [LM Studio](https://lmstudio.ai/)
<|assistant|>
```
## What's new:
Fix tokenizer
## Download a file (not the whole branch) from below:
| Filename | Quant type | File Size | Split | Description |
| -------- | ---------- | --------- | ----- | ----------- |
| [Falcon3-10B-Instruct-f32.gguf](https://huggingface.co/bartowski/Falcon3-10B-Instruct-GGUF/blob/main/Falcon3-10B-Instruct-f32.gguf) | f32 | 41.23GB | false | Full F32 weights. |
| [Falcon3-10B-Instruct-f16.gguf](https://huggingface.co/bartowski/Falcon3-10B-Instruct-GGUF/blob/main/Falcon3-10B-Instruct-f16.gguf) | f16 | 20.62GB | false | Full F16 weights. |
| [Falcon3-10B-Instruct-Q8_0.gguf](https://huggingface.co/bartowski/Falcon3-10B-Instruct-GGUF/blob/main/Falcon3-10B-Instruct-Q8_0.gguf) | Q8_0 | 10.96GB | false | Extremely high quality, generally unneeded but max available quant. |
| [Falcon3-10B-Instruct-Q6_K_L.gguf](https://huggingface.co/bartowski/Falcon3-10B-Instruct-GGUF/blob/main/Falcon3-10B-Instruct-Q6_K_L.gguf) | Q6_K_L | 8.65GB | false | Uses Q8_0 for embed and output weights. Very high quality, near perfect, *recommended*. |