Upload folder using huggingface_hub

This commit is contained in:
Dan Clipca
2025-03-01 18:04:09 +00:00
committed by system
parent e3a2909971
commit a4f3d1e305
4 changed files with 54 additions and 0 deletions

2
.gitattributes vendored
View File

@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
*.zip filter=lfs diff=lfs merge=lfs -text
*.zst filter=lfs diff=lfs merge=lfs -text
*tfevents* filter=lfs diff=lfs merge=lfs -text
Mistral-Small-24B-Instruct-2501-abliterated.imatrix.dat filter=lfs diff=lfs merge=lfs -text
mistral-small-24b-instruct-2501-abliterated-i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:ded897598399b39e468c9d09df0f0a956b5602a9c9de1f60a55e0f496dfffc4e
size 10003529

46
README.md Normal file
View File

@@ -0,0 +1,46 @@
---
base_model: huihui-ai/Mistral-Small-24B-Instruct-2501-abliterated
language:
- en
license: mit
quantized_by: SpongeQuant
tags:
- SpongeQuant
- i1-GGUF
---
Quantized to `i1-GGUF` using [SpongeQuant](https://github.com/SpongeEngine/SpongeQuant), the Oobabooga of LLM quantization.
<div style="display: flex; gap: 20px; align-items: center; margin-top:0;">
<a href="https://github.com/SpongeEngine/SpongeQuant">
<img src="https://huggingface.co/spaces/SpongeEngine/README/resolve/main/github-button.png" width="173">
</a>
<a href="https://discord.gg/azNmr2Gdgy">
<img src="https://huggingface.co/spaces/SpongeEngine/README/resolve/main/discord-button.png" width="173">
</a>
</div>
***
<figure>
<img src="https://huggingface.co/spaces/SpongeEngine/README/resolve/main/110.png" alt="Radio telescope (Arecibo)">
<figcaption>Radio telescope (Arecibo)</figcaption>
</figure>
<figure>
<audio controls>
<source src="https://huggingface.co/spaces/SpongeEngine/README/resolve/main/003.mp3" type="audio/mp3">
Your browser does not support the audio element.
</audio>
<figcaption>Queen & David Bowie Under Pressure (UK, 1981)</figcaption>
</figure>
***
### What is a GGUF?
GGUF is a file format used for running large language models (LLMs) on different types of computers. It supports both regular processors (CPUs) and graphics cards (GPUs), making it easier to run models across a wide range of hardware. Many LLMs require powerful and expensive GPUs, but GGUF improves compatibility and efficiency by optimizing how models are loaded and executed. If a GPU doesn't have enough memory, GGUF can offload parts of the model to the CPU, allowing it to run even when GPU resources are limited. GGUF is designed to work well with quantized models, which use less memory and run faster, making them ideal for lower-end hardware. However, it can also store full-precision models when needed. Thanks to these optimizations, GGUF allows LLMs to run efficiently on everything from high-end GPUs to laptops and even CPU-only systems.
### What is an i1-GGUF?
i1-GGUF is an enhanced type of GGUF model that uses imatrix quantization—a smarter way of reducing model size while preserving key details. Instead of shrinking everything equally, it analyzes the importance of different model components and keeps the most crucial parts more accurate. Like standard GGUF, i1-GGUF allows LLMs to run on various hardware, including CPUs and lower-end GPUs. However, because it prioritizes important weights, i1-GGUF models deliver better responses than traditional GGUF models while maintaining efficiency.

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:f8b6eb710f1eb50edfb600efc9ba4e4a8e4d58c987a00629c5fe57d88f7797cc
size 5273721856