Model: CelesteImperia/Llama-3.2-1B-Instruct-Platinum-GGUF Source: Original Platform
base_model, library_name, pipeline_tag, license, tags
| base_model | library_name | pipeline_tag | license | tags | ||||
|---|---|---|---|---|---|---|---|---|
| meta-llama/Llama-3.2-1B-Instruct | gguf | text-generation | llama3.2 |
|
Llama-3.2-1B-Instruct-GGUF (Platinum Series)
This repository contains the Platinum Series universal GGUF release of Llama-3.2-1B-Instruct. This collection provides multiple quantization levels optimized for cross-platform performance, from mobile devices to high-VRAM workstations.
📦 Available Files & Quantization Details
| File Name | Quantization | Size | Accuracy | Recommended For |
|---|---|---|---|---|
| Llama-3.2-1B-Instruct-Platinum-F16.gguf | FP16 | ~2.5 GB | 100% | Master Reference / Benchmarking |
| Llama-3.2-1B-Instruct-Platinum-Q8_0.gguf | Q8_0 | ~1.3 GB | 99.9% | Platinum Reference / High-Fidelity |
| Llama-3.2-1B-Instruct-Platinum-Q6_K.gguf | Q6_K | ~1.0 GB | 99.7% | High-Quality Inference |
| Llama-3.2-1B-Instruct-Platinum-Q5_K_M.gguf | Q5_K_M | ~0.9 GB | 99.2% | Balanced Desktop Performance |
| Llama-3.2-1B-Instruct-Platinum-Q4_K_M.gguf | Q4_K_M | ~0.7 GB | 98.5% | Mobile / Low-Power Efficiency |
🐍 Python Inference (llama-cpp-python)
To run these engines using Python:
from llama_cpp import Llama
llm = Llama(
model_path="Llama-3.2-1B-Instruct-Platinum-Q8_0.gguf",
n_gpu_layers=-1, # Target all layers to NVIDIA/Apple GPU
n_ctx=4096
)
output = llm("Explain the difference between a class and a struct in C#.", max_tokens=150)
print(output["choices"][0]["text"])
💻 For C# / .NET Users (LLamaSharp)
This collection is fully compatible with .NET applications via the LLamaSharp library.
using LLama.Common;
using LLama;
var parameters = new ModelParams("Llama-3.2-1B-Instruct-Platinum-Q8_0.gguf") {
ContextSize = 4096,
GpuLayerCount = 35
};
using var model = LLamaWeights.LoadFromFile(parameters);
using var context = model.CreateContext(parameters);
var executor = new InteractiveExecutor(context);
Console.WriteLine("Universal Engine Active.");
🏗️ Technical Details
- Optimization Tool: llama.cpp (CUDA-accelerated)
- Architecture: Llama 3.2 (1B)
- Hardware Validation: Dual-GPU (RTX 3090 + RTX A4000)
☕ Support the Forge
| Platform | Support Link |
|---|---|
| Global & India | Support via Razorpay |
Scan to support via UPI (India Only):
Connect with the architect: Abhishek Jaiswal on LinkedIn
Description
Languages
Pip Requirements
100%