From d51a2952e1c9a228a1553e54881c28d8da3c04bb Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 21 Apr 2026 17:36:26 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: ysn-rfd/calme-3.3-instruct-3b-GGUF Source: Original Platform --- .gitattributes | 38 +++++++ README.md | 191 ++++++++++++++++++++++++++++++++ calme-3.3-instruct-3b-q4_0.gguf | 3 + calme-3.3-instruct-3b-q5_0.gguf | 3 + calme-3.3-instruct-3b-q8_0.gguf | 3 + 5 files changed, 238 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 calme-3.3-instruct-3b-q4_0.gguf create mode 100644 calme-3.3-instruct-3b-q5_0.gguf create mode 100644 calme-3.3-instruct-3b-q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..4d169ad --- /dev/null +++ b/.gitattributes @@ -0,0 +1,38 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +calme-3.3-instruct-3b-q8_0.gguf filter=lfs diff=lfs merge=lfs -text +calme-3.3-instruct-3b-q5_0.gguf filter=lfs diff=lfs merge=lfs -text +calme-3.3-instruct-3b-q4_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..615735d --- /dev/null +++ b/README.md @@ -0,0 +1,191 @@ +--- +base_model: MaziyarPanahi/calme-3.3-instruct-3b +datasets: +- MaziyarPanahi/french_instruct_sharegpt +- arcee-ai/EvolKit-20k +language: +- fr +- en +library_name: transformers +license: other +license_name: qwen-research +license_link: https://huggingface.co/Qwen/Qwen2.5-3B/blob/main/LICENSE +pipeline_tag: text-generation +tags: +- chat +- qwen +- qwen2.5 +- finetune +- french +- english +- llama-cpp +- matrixportal +inference: false +model_creator: MaziyarPanahi +quantized_by: MaziyarPanahi +model-index: +- name: calme-3.3-instruct-3b + results: + - task: + type: text-generation + name: Text Generation + dataset: + name: IFEval (0-Shot) + type: HuggingFaceH4/ifeval + args: + num_few_shot: 0 + metrics: + - type: inst_level_strict_acc and prompt_level_strict_acc + value: 64.23 + name: strict accuracy + source: + url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-3.3-instruct-3b + name: Open LLM Leaderboard + - task: + type: text-generation + name: Text Generation + dataset: + name: BBH (3-Shot) + type: BBH + args: + num_few_shot: 3 + metrics: + - type: acc_norm + value: 25.68 + name: normalized accuracy + source: + url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-3.3-instruct-3b + name: Open LLM Leaderboard + - task: + type: text-generation + name: Text Generation + dataset: + name: MATH Lvl 5 (4-Shot) + type: hendrycks/competition_math + args: + num_few_shot: 4 + metrics: + - type: exact_match + value: 0.0 + name: exact match + source: + url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-3.3-instruct-3b + name: Open LLM Leaderboard + - task: + type: text-generation + name: Text Generation + dataset: + name: GPQA (0-shot) + type: Idavidrein/gpqa + args: + num_few_shot: 0 + metrics: + - type: acc_norm + value: 4.36 + name: acc_norm + source: + url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-3.3-instruct-3b + name: Open LLM Leaderboard + - task: + type: text-generation + name: Text Generation + dataset: + name: MuSR (0-shot) + type: TAUR-Lab/MuSR + args: + num_few_shot: 0 + metrics: + - type: acc_norm + value: 9.4 + name: acc_norm + source: + url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-3.3-instruct-3b + name: Open LLM Leaderboard + - task: + type: text-generation + name: Text Generation + dataset: + name: MMLU-PRO (5-shot) + type: TIGER-Lab/MMLU-Pro + config: main + split: test + args: + num_few_shot: 5 + metrics: + - type: acc + value: 25.62 + name: accuracy + source: + url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-3.3-instruct-3b + name: Open LLM Leaderboard +--- + +# ysn-rfd/calme-3.3-instruct-3b-GGUF + This model was converted to GGUF format from [`MaziyarPanahi/calme-3.3-instruct-3b`](https://huggingface.co/MaziyarPanahi/calme-3.3-instruct-3b) using llama.cpp via the ggml.ai's [all-gguf-same-where](https://huggingface.co/spaces/matrixportal/all-gguf-same-where) space. +Refer to the [original model card](https://huggingface.co/MaziyarPanahi/calme-3.3-instruct-3b) for more details on the model. + +## ✅ Quantized Models Download List + +### 🔍 Recommended Quantizations +- **✨ General CPU Use:** [`Q4_K_M`](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q4_k_m.gguf) (Best balance of speed/quality) +- **📱 ARM Devices:** [`Q4_0`](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q4_0.gguf) (Optimized for ARM CPUs) +- **🏆 Maximum Quality:** [`Q8_0`](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q8_0.gguf) (Near-original quality) + +### 📦 Full Quantization Options +| 🚀 Download | 🔢 Type | 📝 Notes | +|:---------|:-----|:------| +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q2_k.gguf) | ![Q2_K](https://img.shields.io/badge/Q2_K-1A73E8) | Basic quantization | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q3_k_s.gguf) | ![Q3_K_S](https://img.shields.io/badge/Q3_K_S-34A853) | Small size | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q3_k_m.gguf) | ![Q3_K_M](https://img.shields.io/badge/Q3_K_M-FBBC05) | Balanced quality | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q3_k_l.gguf) | ![Q3_K_L](https://img.shields.io/badge/Q3_K_L-4285F4) | Better quality | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q4_0.gguf) | ![Q4_0](https://img.shields.io/badge/Q4_0-EA4335) | Fast on ARM | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q4_k_s.gguf) | ![Q4_K_S](https://img.shields.io/badge/Q4_K_S-673AB7) | Fast, recommended | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q4_k_m.gguf) | ![Q4_K_M](https://img.shields.io/badge/Q4_K_M-673AB7) ⭐ | Best balance | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q5_0.gguf) | ![Q5_0](https://img.shields.io/badge/Q5_0-FF6D01) | Good quality | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q5_k_s.gguf) | ![Q5_K_S](https://img.shields.io/badge/Q5_K_S-0F9D58) | Balanced | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q5_k_m.gguf) | ![Q5_K_M](https://img.shields.io/badge/Q5_K_M-0F9D58) | High quality | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q6_k.gguf) | ![Q6_K](https://img.shields.io/badge/Q6_K-4285F4) 🏆 | Very good quality | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-q8_0.gguf) | ![Q8_0](https://img.shields.io/badge/Q8_0-EA4335) ⚡ | Fast, best quality | +| [Download](https://huggingface.co/ysn-rfd/calme-3.3-instruct-3b-GGUF/resolve/main/calme-3.3-instruct-3b-f16.gguf) | ![F16](https://img.shields.io/badge/F16-000000) | Maximum accuracy | + +💡 **Tip:** Use `F16` for maximum precision when quality is critical + + +--- +# 🚀 Applications and Tools for Locally Quantized LLMs +## 🖥️ Desktop Applications + +| Application | Description | Download Link | +|-----------------|----------------------------------------------------------------------------------------------|--------------------------------------------------------------------------------| +| **Llama.cpp** | A fast and efficient inference engine for GGUF models. | [GitHub Repository](https://github.com/ggml-org/llama.cpp) | +| **Ollama** | A streamlined solution for running LLMs locally. | [Website](https://ollama.com/) | +| **AnythingLLM** | An AI-powered knowledge management tool. | [GitHub Repository](https://github.com/Mintplex-Labs/anything-llm) | +| **Open WebUI** | A user-friendly web interface for running local LLMs. | [GitHub Repository](https://github.com/open-webui/open-webui) | +| **GPT4All** | A user-friendly desktop application supporting various LLMs, compatible with GGUF models. | [GitHub Repository](https://github.com/nomic-ai/gpt4all) | +| **LM Studio** | A desktop application designed to run and manage local LLMs, supporting GGUF format. | [Website](https://lmstudio.ai/) | +| **GPT4All Chat**| A chat application compatible with GGUF models for local, offline interactions. | [GitHub Repository](https://github.com/nomic-ai/gpt4all) | + +--- + +## 📱 Mobile Applications + +| Application | Description | Download Link | +|-------------------|----------------------------------------------------------------------------------------------|--------------------------------------------------------------------------------| +| **ChatterUI** | A simple and lightweight LLM app for mobile devices. | [GitHub Repository](https://github.com/Vali-98/ChatterUI) | +| **Maid** | Mobile Artificial Intelligence Distribution for running AI models on mobile devices. | [GitHub Repository](https://github.com/Mobile-Artificial-Intelligence/maid) | +| **PocketPal AI** | A mobile AI assistant powered by local models. | [GitHub Repository](https://github.com/a-ghorbani/pocketpal-ai) | +| **Layla** | A flexible platform for running various AI models on mobile devices. | [Website](https://www.layla-network.ai/) | + +--- + +## 🎨 Image Generation Applications + +| Application | Description | Download Link | +|-------------------------------------|----------------------------------------------------------------------------------------------|--------------------------------------------------------------------------------| +| **Stable Diffusion** | An open-source AI model for generating images from text. | [GitHub Repository](https://github.com/CompVis/stable-diffusion) | +| **Stable Diffusion WebUI** | A web application providing access to Stable Diffusion models via a browser interface. | [GitHub Repository](https://github.com/AUTOMATIC1111/stable-diffusion-webui) | +| **Local Dream** | Android Stable Diffusion with Snapdragon NPU acceleration. Also supports CPU inference. | [GitHub Repository](https://github.com/xororz/local-dream) | +| **Stable-Diffusion-Android (SDAI)** | An open-source AI art application for Android devices, enabling digital art creation. | [GitHub Repository](https://github.com/ShiftHackZ/Stable-Diffusion-Android) | + +--- + diff --git a/calme-3.3-instruct-3b-q4_0.gguf b/calme-3.3-instruct-3b-q4_0.gguf new file mode 100644 index 0000000..e642e10 --- /dev/null +++ b/calme-3.3-instruct-3b-q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a8d8a6747c846272bad91fd1692cdedd6054372be854ccc832b70759be4f145 +size 1822850592 diff --git a/calme-3.3-instruct-3b-q5_0.gguf b/calme-3.3-instruct-3b-q5_0.gguf new file mode 100644 index 0000000..a8c856e --- /dev/null +++ b/calme-3.3-instruct-3b-q5_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3827ab64bc241efc5ebf092a2f4eb056792f78b48d737d2ba92162f8aec853e5 +size 2169667104 diff --git a/calme-3.3-instruct-3b-q8_0.gguf b/calme-3.3-instruct-3b-q8_0.gguf new file mode 100644 index 0000000..e222f34 --- /dev/null +++ b/calme-3.3-instruct-3b-q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35c96cf1fe1e43328f11d40365d5a577e80078546d6b0917178434d162c13f5d +size 3285476896