Update metadata with huggingface_hub
This commit is contained in:
11
README.md
11
README.md
@@ -1,14 +1,14 @@
|
||||
---
|
||||
quantized_by: bartowski
|
||||
pipeline_tag: text-generation
|
||||
license: apache-2.0
|
||||
base_model_relation: quantized
|
||||
base_model: cognitivecomputations/Dolphin-Mistral-24B-Venice-Edition
|
||||
base_model_relation: quantized
|
||||
license: apache-2.0
|
||||
---
|
||||
|
||||
## Llamacpp imatrix Quantizations of Dolphin-Mistral-24B-Venice-Edition by cognitivecomputations
|
||||
|
||||
Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b5317">b5317</a> for quantization.
|
||||
Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b5835">b5835</a> for quantization.
|
||||
|
||||
Original model: https://huggingface.co/cognitivecomputations/Dolphin-Mistral-24B-Venice-Edition
|
||||
|
||||
@@ -24,6 +24,10 @@ Run them directly with [llama.cpp](https://github.com/ggerganov/llama.cpp), or a
|
||||
<s>[SYSTEM_PROMPT]{system_prompt}[/SYSTEM_PROMPT][INST]{prompt}[/INST]
|
||||
```
|
||||
|
||||
## What's new:
|
||||
|
||||
Original model updated
|
||||
|
||||
## Download a file (not the whole branch) from below:
|
||||
|
||||
| Filename | Quant type | File Size | Split | Description |
|
||||
@@ -54,6 +58,7 @@ Run them directly with [llama.cpp](https://github.com/ggerganov/llama.cpp), or a
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-IQ2_M.gguf](https://huggingface.co/bartowski/cognitivecomputations_Dolphin-Mistral-24B-Venice-Edition-GGUF/blob/main/cognitivecomputations_Dolphin-Mistral-24B-Venice-Edition-IQ2_M.gguf) | IQ2_M | 8.11GB | false | Relatively low quality, uses SOTA techniques to be surprisingly usable. |
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-IQ2_S.gguf](https://huggingface.co/bartowski/cognitivecomputations_Dolphin-Mistral-24B-Venice-Edition-GGUF/blob/main/cognitivecomputations_Dolphin-Mistral-24B-Venice-Edition-IQ2_S.gguf) | IQ2_S | 7.48GB | false | Low quality, uses SOTA techniques to be usable. |
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-IQ2_XS.gguf](https://huggingface.co/bartowski/cognitivecomputations_Dolphin-Mistral-24B-Venice-Edition-GGUF/blob/main/cognitivecomputations_Dolphin-Mistral-24B-Venice-Edition-IQ2_XS.gguf) | IQ2_XS | 7.21GB | false | Low quality, uses SOTA techniques to be usable. |
|
||||
| [Dolphin-Mistral-24B-Venice-Edition-IQ2_XXS.gguf](https://huggingface.co/bartowski/cognitivecomputations_Dolphin-Mistral-24B-Venice-Edition-GGUF/blob/main/cognitivecomputations_Dolphin-Mistral-24B-Venice-Edition-IQ2_XXS.gguf) | IQ2_XXS | 6.55GB | false | Very low quality, uses SOTA techniques to be usable. |
|
||||
|
||||
## Embed/output weights
|
||||
|
||||
|
||||
Reference in New Issue
Block a user