Update README.md
This commit is contained in:
@@ -132,7 +132,7 @@ The process to generate these models is roughly as follows:
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-pruned-Q5_K_S](./Dolphin-Mistral-24B-Venice-Edition-pruned-Q5_K_S.gguf) | 18.199918 ±0.142513 | 75.20% | 1.160040 ±0.004229 | 36.220 ±0.088 |
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-pruned-Q6_K](./Dolphin-Mistral-24B-Venice-Edition-pruned-Q6_K.gguf) | 18.213825 ±0.142965 | 75.05% | 1.158026 ±0.004262 | 36.219 ±0.088 |
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-pruned-Q8_0](./Dolphin-Mistral-24B-Venice-Edition-pruned-Q8_0.gguf) | 18.203515 ±0.142826 | 75.02% | 1.158351 ±0.004265 | 36.227 ±0.088 |
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-pruned-F16](./Dolphin-Mistral-24B-Venice-Edition-pruned-F16.gguf) | 6.180577 ±0.041038 | 100% | N/A | N/A |
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-pruned-F16](./Dolphin-Mistral-24B-Venice-Edition-F16.gguf) | 6.180577 ±0.041038 | 100% | N/A | N/A |
|
||||
|
||||
### ARC, HellaSwag, MMLU, Truthful QA and WinoGrande scores
|
||||
Scores generated using [llama-perplexity][ppl] with 750 tasks per test, and a context size of 768 tokens.
|
||||
@@ -154,7 +154,7 @@ For the test data used in the generation of these scores, follow the appropriate
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-pruned-Q5_K_S](./Dolphin-Mistral-24B-Venice-Edition-pruned-Q5_K_S.gguf) | 67.3333 ±1.7137 | 81.47 | 44.2667 ±1.8149 | 38.6667 ±1.7794 | 74.2667 ±1.5974 | 61.20 |
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-pruned-Q6_K](./Dolphin-Mistral-24B-Venice-Edition-pruned-Q6_K.gguf) | 67.4667 ±1.7119 | 81.07 | 44.5333 ±1.8160 | 39.6000 ±1.7870 | 73.8667 ±1.6054 | 61.31 |
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-pruned-Q8_0](./Dolphin-Mistral-24B-Venice-Edition-pruned-Q8_0.gguf) | 68.1333 ±1.7026 | 81.33 | 44.9333 ±1.8176 | 38.2667 ±1.7759 | 74.4000 ±1.5947 | 61.41 |
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-pruned-F16](./Dolphin-Mistral-24B-Venice-Edition-pruned-F16.gguf) | 70.8000 ±1.6614 | 84.53 | 45.3333 ±1.8190 | 38.1333 ±1.7748 | 80.2667 ±1.4542 | 63.81 |
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-pruned-F16](./Dolphin-Mistral-24B-Venice-Edition-F16.gguf) | 70.8000 ±1.6614 | 84.53 | 45.3333 ±1.8190 | 38.1333 ±1.7748 | 80.2667 ±1.4542 | 63.81 |
|
||||
|
||||
### Tokens per Second - Benchmarks
|
||||
Scores generated using [llama-bench][bch]. Naive (`llama-quantize` with no optimization) Q4_K_M quantization included for comparison.
|
||||
|
||||
Reference in New Issue
Block a user