From 853c4317c0c4c129a8f75c36cf711293a3fdaf40 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 21 Apr 2026 17:42:48 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: EasierAI/Qwen-2.5-0.5B Source: Original Platform --- .gitattributes | 43 +++++++++++++++++++++++++ Qwen-2.5-0.5B-Instruct-Q3_K_L.gguf | 3 ++ Qwen-2.5-0.5B-Instruct-Q4_K_M.gguf | 3 ++ Qwen-2.5-0.5B-Instruct-Q6_K.gguf | 3 ++ Qwen-2.5-0.5B-Instruct-Q8_0.gguf | 3 ++ README.md | 50 ++++++++++++++++++++++++++++++ 6 files changed, 105 insertions(+) create mode 100644 .gitattributes create mode 100644 Qwen-2.5-0.5B-Instruct-Q3_K_L.gguf create mode 100644 Qwen-2.5-0.5B-Instruct-Q4_K_M.gguf create mode 100644 Qwen-2.5-0.5B-Instruct-Q6_K.gguf create mode 100644 Qwen-2.5-0.5B-Instruct-Q8_0.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..339513a --- /dev/null +++ b/.gitattributes @@ -0,0 +1,43 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Qwen2.5-0.5B-Instruct-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Qwen2.5-0.5B-Instruct-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Qwen2.5-0.5B-Instruct-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Qwen2.5-0.5B-Instruct-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-2.5-0.5B-Instruct-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-2.5-0.5B-Instruct-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-2.5-0.5B-Instruct-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-2.5-0.5B-Instruct-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Qwen-2.5-0.5B-Instruct-Q3_K_L.gguf b/Qwen-2.5-0.5B-Instruct-Q3_K_L.gguf new file mode 100644 index 0000000..93d8884 --- /dev/null +++ b/Qwen-2.5-0.5B-Instruct-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b36b5c87047351aae5964ff9cbcd04b68a0da949618965c8adfc947da641b1f +size 369358144 diff --git a/Qwen-2.5-0.5B-Instruct-Q4_K_M.gguf b/Qwen-2.5-0.5B-Instruct-Q4_K_M.gguf new file mode 100644 index 0000000..3d5f0b4 --- /dev/null +++ b/Qwen-2.5-0.5B-Instruct-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa4d41b65761ed565cac6b5f62e35135d050408b033114a128ab308c02b2e83a +size 397807936 diff --git a/Qwen-2.5-0.5B-Instruct-Q6_K.gguf b/Qwen-2.5-0.5B-Instruct-Q6_K.gguf new file mode 100644 index 0000000..f287990 --- /dev/null +++ b/Qwen-2.5-0.5B-Instruct-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d1e192b0fe6f4cde8696836a4b5d77bb24550988aab03146c16089d6e7385ed +size 505736512 diff --git a/Qwen-2.5-0.5B-Instruct-Q8_0.gguf b/Qwen-2.5-0.5B-Instruct-Q8_0.gguf new file mode 100644 index 0000000..61de0d5 --- /dev/null +++ b/Qwen-2.5-0.5B-Instruct-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d96e30253a76970864323987e3e77eeaf170312952fd61d18d343c25cfd6e52 +size 531068224 diff --git a/README.md b/README.md new file mode 100644 index 0000000..4f68418 --- /dev/null +++ b/README.md @@ -0,0 +1,50 @@ +--- +base_model: Qwen/Qwen2.5-0.5B-Instruct +language: +- zho +- eng +- fra +- spa +- por +- deu +- ita +- rus +- jpn +- kor +- vie +- tha +- ara +license: apache-2.0 +license_link: https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct/blob/main/LICENSE +pipeline_tag: text-generation +tags: +- chat +quantized_by: bartowski +--- +## 💫 Community Model> Qwen2.5 0.5B Instruct by Qwen + +*👾 [LM Studio](https://lmstudio.ai) Community models highlights program. Highlighting new & noteworthy models by the community. Join the conversation on [Discord](https://discord.gg/aPQfnNkxGC)*. + +**Model creator:** [Qwen](https://huggingface.co/Qwen)
+**Original model**: [Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct)
+**GGUF quantization:** provided by [bartowski](https://huggingface.co/bartowski) based on `llama.cpp` release [b3772](https://github.com/ggerganov/llama.cpp/releases/tag/b3772)
+ +## Technical Details + +Long context: Support for 32k tokens and 8k token generation + +Large-scale training dataset: Encompasses a huge range of knowledge. + +Enhanced structured data understanding and generation. + +Over 29 languages including Chinese, English, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, and Arabic. + +More details available [here](https://qwenlm.github.io/blog/qwen2.5/) + +## Special thanks + +🙏 Special thanks to [Georgi Gerganov](https://github.com/ggerganov) and the whole team working on [llama.cpp](https://github.com/ggerganov/llama.cpp/) for making all of this possible. + +## Disclaimers + +LM Studio is not the creator, originator, or owner of any Model featured in the Community Model Program. Each Community Model is created and provided by third parties. LM Studio does not endorse, support, represent or guarantee the completeness, truthfulness, accuracy, or reliability of any Community Model. You understand that Community Models can produce content that might be offensive, harmful, inaccurate or otherwise inappropriate, or deceptive. Each Community Model is the sole responsibility of the person or entity who originated such Model. LM Studio may not monitor or control the Community Models and cannot, and does not, take responsibility for any such Model. LM Studio disclaims all warranties or guarantees about the accuracy, reliability or benefits of the Community Models. LM Studio further disclaims any warranty that the Community Model will meet your requirements, be secure, uninterrupted or available at any time or location, or error-free, viruses-free, or that any errors will be corrected, or otherwise. You will be solely responsible for any damage resulting from your use of or access to the Community Models, your downloading of any Community Model, or use of any other Community Model provided by or through LM Studio.