From e51a174030deab5dbf022d4368eeea164d229a9e Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Wed, 6 May 2026 01:40:38 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: CelesteImperia/Llama-3.2-3B-Instruct-Platinum-GGUF Source: Original Platform --- .gitattributes | 40 +++++++++ LICENSE | 6 ++ Llama-3.2-3B-Instruct-Platinum-F16.gguf | 3 + Llama-3.2-3B-Instruct-Platinum-Q4_K_M.gguf | 3 + Llama-3.2-3B-Instruct-Platinum-Q5_K_M.gguf | 3 + Llama-3.2-3B-Instruct-Platinum-Q6_K.gguf | 3 + Llama-3.2-3B-Instruct-Platinum-Q8_0.gguf | 3 + README.md | 96 ++++++++++++++++++++++ requirements.txt | 4 + 9 files changed, 161 insertions(+) create mode 100644 .gitattributes create mode 100644 LICENSE create mode 100644 Llama-3.2-3B-Instruct-Platinum-F16.gguf create mode 100644 Llama-3.2-3B-Instruct-Platinum-Q4_K_M.gguf create mode 100644 Llama-3.2-3B-Instruct-Platinum-Q5_K_M.gguf create mode 100644 Llama-3.2-3B-Instruct-Platinum-Q6_K.gguf create mode 100644 Llama-3.2-3B-Instruct-Platinum-Q8_0.gguf create mode 100644 README.md create mode 100644 requirements.txt diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..538f0c0 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,40 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Llama-3.2-3B-Instruct-Platinum-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-3B-Instruct-Platinum-F16.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-3B-Instruct-Platinum-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-3B-Instruct-Platinum-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-3B-Instruct-Platinum-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/LICENSE b/LICENSE new file mode 100644 index 0000000..bd18cce --- /dev/null +++ b/LICENSE @@ -0,0 +1,6 @@ +Llama 3.2 Community License Agreement + +This model is subject to the Llama 3.2 Community License. +The full license agreement, acceptable use policy, and +redistribution terms are available at: +https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE \ No newline at end of file diff --git a/Llama-3.2-3B-Instruct-Platinum-F16.gguf b/Llama-3.2-3B-Instruct-Platinum-F16.gguf new file mode 100644 index 0000000..a426f72 --- /dev/null +++ b/Llama-3.2-3B-Instruct-Platinum-F16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f15c41c9a5fd3b23f2dead5b540356f5edbd9d1bcc3547110ac875a56d7b61b9 +size 6433688000 diff --git a/Llama-3.2-3B-Instruct-Platinum-Q4_K_M.gguf b/Llama-3.2-3B-Instruct-Platinum-Q4_K_M.gguf new file mode 100644 index 0000000..a0bdf9e --- /dev/null +++ b/Llama-3.2-3B-Instruct-Platinum-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2b542d8ed920319e862956086fe6867b556d46930ae55740ee7613df25a50b4 +size 2019377600 diff --git a/Llama-3.2-3B-Instruct-Platinum-Q5_K_M.gguf b/Llama-3.2-3B-Instruct-Platinum-Q5_K_M.gguf new file mode 100644 index 0000000..8cd6cce --- /dev/null +++ b/Llama-3.2-3B-Instruct-Platinum-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9b4e4e6a5057589a8c3d16e9e1ecd83a72fd6dd9730e93bb1f2f99b27ab4ca3 +size 2322153920 diff --git a/Llama-3.2-3B-Instruct-Platinum-Q6_K.gguf b/Llama-3.2-3B-Instruct-Platinum-Q6_K.gguf new file mode 100644 index 0000000..65712fa --- /dev/null +++ b/Llama-3.2-3B-Instruct-Platinum-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7f7c50bdf5cebba0e7ca1edd0ec2b7468fd4e3caf633817c502774540eea7f7 +size 2643853760 diff --git a/Llama-3.2-3B-Instruct-Platinum-Q8_0.gguf b/Llama-3.2-3B-Instruct-Platinum-Q8_0.gguf new file mode 100644 index 0000000..6a675e4 --- /dev/null +++ b/Llama-3.2-3B-Instruct-Platinum-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4840c99335d9be74ce1718659e98380224a3d7384761a8dc305847cf440a6c3 +size 3421899200 diff --git a/README.md b/README.md new file mode 100644 index 0000000..ee12dbd --- /dev/null +++ b/README.md @@ -0,0 +1,96 @@ +--- +base_model: meta-llama/Llama-3.2-3B-Instruct +library_name: gguf +pipeline_tag: text-generation +license: llama3.2 +tags: +- gguf +- llama-cpp +- llama-3.2 +- celeste-imperia +--- + +# Llama-3.2-3B-Instruct-GGUF (Platinum Series) + +![Status](https://img.shields.io/badge/Status-Active-success) +![Format](https://img.shields.io/badge/Format-GGUF-green) +![Series](https://img.shields.io/badge/Series-Platinum-silver) +[![Support](https://img.shields.io/badge/Support-Razorpay-orange)](https://razorpay.me/@huggingface) + +This repository contains the **Platinum Series** universal GGUF release of **Llama-3.2-3B-Instruct**. This collection provides multiple quantization levels optimized for cross-platform performance, offering a significant reasoning upgrade over the 1B variant while maintaining exceptional speed on consumer hardware. + +## 📦 Available Files & Quantization Details + +| File Name | Quantization | Size | Accuracy | Recommended For | +| :--- | :--- | :--- | :--- | :--- | +| **Llama-3.2-3B-Instruct-Platinum-F16.gguf** | FP16 | ~6.5 GB | 100% | Master Reference / Benchmarking | +| **Llama-3.2-3B-Instruct-Platinum-Q8_0.gguf** | Q8_0 | ~3.4 GB | 99.9% | Platinum Reference / High-Fidelity | +| **Llama-3.2-3B-Instruct-Platinum-Q6_K.gguf** | Q6_K | ~2.7 GB | 99.7% | High-Quality Reasoning | +| **Llama-3.2-3B-Instruct-Platinum-Q5_K_M.gguf** | Q5_K_M | ~2.4 GB | 99.3% | Balanced Desktop Performance | +| **Llama-3.2-3B-Instruct-Platinum-Q4_K_M.gguf** | Q4_K_M | ~2.0 GB | 98.6% | Edge Devices / Efficiency | + +--- + +## 🐍 Python Inference (llama-cpp-python) + +To run these engines using Python: + +```python +from llama_cpp import Llama + +llm = Llama( + model_path="Llama-3.2-3B-Instruct-Platinum-Q8_0.gguf", + n_gpu_layers=-1, # Target all layers to NVIDIA/Apple GPU + n_ctx=4096 +) + +output = llm("Discuss the architectural benefits of Llama 3.2 3B.", max_tokens=200) +print(output["choices"][0]["text"]) +``` + +--- + +## 💻 For C# / .NET Users (LLamaSharp) + +This collection is fully compatible with .NET applications via the ``LLamaSharp`` library. + +```csharp +using LLama.Common; +using LLama; + +var parameters = new ModelParams("Llama-3.2-3B-Instruct-Platinum-Q8_0.gguf") { + ContextSize = 4096, + GpuLayerCount = 35 +}; + +using var model = LLamaWeights.LoadFromFile(parameters); +using var context = model.CreateContext(parameters); +var executor = new InteractiveExecutor(context); + +Console.WriteLine("Universal Engine Active."); +``` + +--- + +## 🏗️ Technical Details +- **Optimization Tool:** llama.cpp (CUDA-accelerated) +- **Architecture:** Llama 3.2 (3B) +- **Hardware Validation:** Dual-GPU (RTX 3090 + RTX A4000) + +--- + +### ☕ Support the Forge + +Maintaining the production line for high-fidelity models requires significant hardware resources. If these tools power your research or industrial projects, please consider supporting the development: + +| Platform | Support Link | +| :--- | :--- | +| **Global & India** | [Support via Razorpay](https://razorpay.me/@huggingface) | + +**Scan to support via UPI (India Only):** + + + +--- + +**Connect with the architect:** [Abhishek Jaiswal on LinkedIn](https://www.linkedin.com/in/abhishek-jaiswal-524056a/) \ No newline at end of file diff --git a/requirements.txt b/requirements.txt new file mode 100644 index 0000000..86c375f --- /dev/null +++ b/requirements.txt @@ -0,0 +1,4 @@ +optimum-intel[openvino,nncf]>=1.20.0 +transformers>=4.45.0 +accelerate +sentencepiece \ No newline at end of file