--- base_model: meta-llama/Llama-3.2-1B-Instruct library_name: gguf pipeline_tag: text-generation license: llama3.2 tags: - gguf - llama-cpp - llama-3.2 - celeste-imperia --- # Llama-3.2-1B-Instruct-GGUF (Platinum Series) ![Status](https://img.shields.io/badge/Status-Active-success) ![Format](https://img.shields.io/badge/Format-GGUF-green) ![Series](https://img.shields.io/badge/Series-Platinum-silver) [![Support](https://img.shields.io/badge/Support-Razorpay-orange)](https://razorpay.me/@huggingface) This repository contains the **Platinum Series** universal GGUF release of **Llama-3.2-1B-Instruct**. This collection provides multiple quantization levels optimized for cross-platform performance, from mobile devices to high-VRAM workstations. ## 📦 Available Files & Quantization Details | File Name | Quantization | Size | Accuracy | Recommended For | | :--- | :--- | :--- | :--- | :--- | | **Llama-3.2-1B-Instruct-Platinum-F16.gguf** | FP16 | ~2.5 GB | 100% | Master Reference / Benchmarking | | **Llama-3.2-1B-Instruct-Platinum-Q8_0.gguf** | Q8_0 | ~1.3 GB | 99.9% | Platinum Reference / High-Fidelity | | **Llama-3.2-1B-Instruct-Platinum-Q6_K.gguf** | Q6_K | ~1.0 GB | 99.7% | High-Quality Inference | | **Llama-3.2-1B-Instruct-Platinum-Q5_K_M.gguf** | Q5_K_M | ~0.9 GB | 99.2% | Balanced Desktop Performance | | **Llama-3.2-1B-Instruct-Platinum-Q4_K_M.gguf** | Q4_K_M | ~0.7 GB | 98.5% | Mobile / Low-Power Efficiency | --- ## 🐍 Python Inference (llama-cpp-python) To run these engines using Python: ```python from llama_cpp import Llama llm = Llama( model_path="Llama-3.2-1B-Instruct-Platinum-Q8_0.gguf", n_gpu_layers=-1, # Target all layers to NVIDIA/Apple GPU n_ctx=4096 ) output = llm("Explain the difference between a class and a struct in C#.", max_tokens=150) print(output["choices"][0]["text"]) ``` --- ## 💻 For C# / .NET Users (LLamaSharp) This collection is fully compatible with .NET applications via the ``LLamaSharp`` library. ```csharp using LLama.Common; using LLama; var parameters = new ModelParams("Llama-3.2-1B-Instruct-Platinum-Q8_0.gguf") { ContextSize = 4096, GpuLayerCount = 35 }; using var model = LLamaWeights.LoadFromFile(parameters); using var context = model.CreateContext(parameters); var executor = new InteractiveExecutor(context); Console.WriteLine("Universal Engine Active."); ``` --- ## 🏗️ Technical Details - **Optimization Tool:** llama.cpp (CUDA-accelerated) - **Architecture:** Llama 3.2 (1B) - **Hardware Validation:** Dual-GPU (RTX 3090 + RTX A4000) --- ### ☕ Support the Forge | Platform | Support Link | | :--- | :--- | | **Global & India** | [Support via Razorpay](https://razorpay.me/@huggingface) | **Scan to support via UPI (India Only):** --- **Connect with the architect:** [Abhishek Jaiswal on LinkedIn](https://www.linkedin.com/in/abhishek-jaiswal-524056a/)