c2072f4d38e9710e7bff3f406a33cd06376380e1
Model: ThijsL202/CraneAILabs_swahili-gemma-1b-GGUF Source: Original Platform
base_model, language, library_name, pipeline_tag, quantized_by, tags
| base_model | language | library_name | pipeline_tag | quantized_by | tags | ||||
|---|---|---|---|---|---|---|---|---|---|
| CraneAILabs/swahili-gemma-1b |
|
transformers | text-generation | ThijsL202 |
|
CraneAILabs_swahili-gemma-1b - GGUF Standard
📊 Quantization Details
- Base Model: CraneAILabs/swahili-gemma-1b
- Quantization: Standard
- Total Size: 4.03 GB (5 files)
📦 Standard Quantizations Classic GGUF quantizations without importance matrix enhancement.
📁 Available Files
| Quantization | Size | Download |
|---|---|---|
| Q2_K | 690MB | ⬇️ |
| Q3_K_L | 752MB | ⬇️ |
| Q4_K_M | 806MB | ⬇️ |
| Q6_K | 1.0GB | ⬇️ |
| Q8_0 | 1.1GB | ⬇️ |
🚀 Quick Start
llama.cpp
./llama-cli -m CraneAILabs_swahili-gemma-1b.Q6_K.gguf -p "Your prompt here" -n 512
Python
from llama_cpp import Llama
llm = Llama(model_path="./CraneAILabs_swahili-gemma-1b.Q6_K.gguf", n_ctx=2048)
output = llm("Your prompt here", max_tokens=512)
print(output["choices"][0]["text"])
📊 Model Information
Original model: CraneAILabs/swahili-gemma-1b
Quantized using llama.cpp
Description