--- base_model: Qwen/Qwen2.5-7B-Instruct library_name: gguf pipeline_tag: text-generation license: apache-2.0 tags: - gguf - llama-cpp - qwen2.5 - celeste-imperia --- # Qwen-2.5-7B-Instruct-GGUF (Platinum Series) ![Status](https://img.shields.io/badge/Status-Active-success) ![Format](https://img.shields.io/badge/Format-GGUF-green) ![Series](https://img.shields.io/badge/Series-Platinum-silver) [![Support](https://img.shields.io/badge/Support-Razorpay-orange)](https://razorpay.me/@huggingface) This repository contains the **Platinum Series** universal GGUF release of **Qwen-2.5-7B-Instruct**. This collection provides multiple quantization levels optimized for cross-platform performance, offering professional-grade reasoning and coding capabilities. ## 📦 Available Files & Quantization Details | File Name | Quantization | Size | Accuracy | Recommended For | | :--- | :--- | :--- | :--- | :--- | | **Qwen2.5-7B-Instruct-Platinum-F16.gguf** | FP16 | ~15.0 GB | 100% | Master Reference / Benchmarking | | **Qwen2.5-7B-Instruct-Platinum-Q8_0.gguf** | Q8_0 | ~8.0 GB | 99.9% | Platinum Reference / High-Fidelity | | **Qwen2.5-7B-Instruct-Platinum-Q6_K.gguf** | Q6_K | ~6.3 GB | 99.8% | High-Quality Reasoning | | **Qwen2.5-7B-Instruct-Platinum-Q5_K_M.gguf** | Q5_K_M | ~5.5 GB | 99.5% | Balanced Desktop Performance | | **Qwen2.5-7B-Instruct-Platinum-Q4_K_M.gguf** | Q4_K_M | ~4.7 GB | 99.0% | Efficiency / Mid-Range Hardware | --- ## 🐍 Python Inference (llama-cpp-python) To run these engines using Python: ```python from llama_cpp import Llama llm = Llama( model_path="Qwen2.5-7B-Instruct-Platinum-Q8_0.gguf", n_gpu_layers=-1, # Target all layers to NVIDIA/Apple GPU n_ctx=4096 ) output = llm("Explain the core improvements in Qwen 2.5.", max_tokens=150) print(output["choices"][0]["text"]) ``` --- ## 💻 For C# / .NET Users (LLamaSharp) This collection is fully compatible with .NET applications via the ``LLamaSharp`` library. ```csharp using LLama.Common; using LLama; var parameters = new ModelParams("Qwen2.5-7B-Instruct-Platinum-Q8_0.gguf") { ContextSize = 4096, GpuLayerCount = 35 }; using var model = LLamaWeights.LoadFromFile(parameters); using var context = model.CreateContext(parameters); var executor = new InteractiveExecutor(context); Console.WriteLine("Universal Engine Active."); ``` --- ## 🏗️ Technical Details - **Optimization Tool:** llama.cpp (CUDA-accelerated) - **Architecture:** Qwen-2.5 (7B) - **Hardware Validation:** Dual-GPU (RTX 3090 + RTX A4000) --- ### ☕ Support the Forge Maintaining the production line for high-fidelity models requires significant hardware resources. If these tools power your research or industrial projects, please consider supporting the development: | Platform | Support Link | | :--- | :--- | | **Global & India** | [Support via Razorpay](https://razorpay.me/@huggingface) | **Scan to support via UPI (India Only):** --- **Connect with the architect:** [Abhishek Jaiswal on LinkedIn](https://www.linkedin.com/in/abhishek-jaiswal-524056a/)