From f7cdad162e6930068e63a6d8040e3c4f308fdb35 Mon Sep 17 00:00:00 2001 From: Dan Clipca Date: Tue, 4 Nov 2025 13:54:43 +0000 Subject: [PATCH] Update README.md --- README.md | 25 ------------------------- 1 file changed, 25 deletions(-) diff --git a/README.md b/README.md index 106a486..440117c 100644 --- a/README.md +++ b/README.md @@ -12,31 +12,6 @@ tags: Quantized to `i1-GGUF` using [SpongeQuant](https://github.com/SpongeEngine/SpongeQuant), the Oobabooga of LLM quantization. -
- - - - - - -
- -*** -
- UN Building Day -
UN Building Day
-
- -
- -
NENA – 99 Luftballons (Germany, 1983)
-
- -*** - ### What is a GGUF? GGUF is a file format used for running large language models (LLMs) on different types of computers. It supports both regular processors (CPUs) and graphics cards (GPUs), making it easier to run models across a wide range of hardware. Many LLMs require powerful and expensive GPUs, but GGUF improves compatibility and efficiency by optimizing how models are loaded and executed. If a GPU doesn't have enough memory, GGUF can offload parts of the model to the CPU, allowing it to run even when GPU resources are limited. GGUF is designed to work well with quantized models, which use less memory and run faster, making them ideal for lower-end hardware. However, it can also store full-precision models when needed. Thanks to these optimizations, GGUF allows LLMs to run efficiently on everything from high-end GPUs to laptops and even CPU-only systems.