From afce2bf02cd852edf18dd1ef9ec7b2e9b0d1ed06 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sat, 18 Apr 2026 07:31:20 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: afrideva/MiniChat-1.5-3B-GGUF Source: Original Platform --- .gitattributes | 42 +++++++++++++++ README.md | 105 ++++++++++++++++++++++++++++++++++++ minichat-1.5-3b.fp16.gguf | 3 ++ minichat-1.5-3b.q2_k.gguf | 3 ++ minichat-1.5-3b.q3_k_m.gguf | 3 ++ minichat-1.5-3b.q4_k_m.gguf | 3 ++ minichat-1.5-3b.q5_k_m.gguf | 3 ++ minichat-1.5-3b.q6_k.gguf | 3 ++ minichat-1.5-3b.q8_0.gguf | 3 ++ 9 files changed, 168 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 minichat-1.5-3b.fp16.gguf create mode 100644 minichat-1.5-3b.q2_k.gguf create mode 100644 minichat-1.5-3b.q3_k_m.gguf create mode 100644 minichat-1.5-3b.q4_k_m.gguf create mode 100644 minichat-1.5-3b.q5_k_m.gguf create mode 100644 minichat-1.5-3b.q6_k.gguf create mode 100644 minichat-1.5-3b.q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..be962b2 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,42 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +minichat-1.5-3b.fp16.gguf filter=lfs diff=lfs merge=lfs -text +minichat-1.5-3b.q2_k.gguf filter=lfs diff=lfs merge=lfs -text +minichat-1.5-3b.q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text +minichat-1.5-3b.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +minichat-1.5-3b.q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text +minichat-1.5-3b.q6_k.gguf filter=lfs diff=lfs merge=lfs -text +minichat-1.5-3b.q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..3e0fa4f --- /dev/null +++ b/README.md @@ -0,0 +1,105 @@ +--- +base_model: GeneZC/MiniChat-1.5-3B +inference: false +language: +- en +- zh +library_name: transformers +license: apache-2.0 +model_creator: GeneZC +model_name: MiniChat-1.5-3B +pipeline_tag: text-generation +quantized_by: afrideva +tags: +- gguf +- ggml +- quantized +- q2_k +- q3_k_m +- q4_k_m +- q5_k_m +- q6_k +- q8_0 +widget: +- text: " [|User|] Hi \U0001F44B [|Assistant|]" +--- +# GeneZC/MiniChat-1.5-3B-GGUF + +Quantized GGUF model files for [MiniChat-1.5-3B](https://huggingface.co/GeneZC/MiniChat-1.5-3B) from [GeneZC](https://huggingface.co/GeneZC) + + +| Name | Quant method | Size | +| ---- | ---- | ---- | +| [minichat-1.5-3b.fp16.gguf](https://huggingface.co/afrideva/MiniChat-1.5-3B-GGUF/resolve/main/minichat-1.5-3b.fp16.gguf) | fp16 | 6.04 GB | +| [minichat-1.5-3b.q2_k.gguf](https://huggingface.co/afrideva/MiniChat-1.5-3B-GGUF/resolve/main/minichat-1.5-3b.q2_k.gguf) | q2_k | 1.30 GB | +| [minichat-1.5-3b.q3_k_m.gguf](https://huggingface.co/afrideva/MiniChat-1.5-3B-GGUF/resolve/main/minichat-1.5-3b.q3_k_m.gguf) | q3_k_m | 1.51 GB | +| [minichat-1.5-3b.q4_k_m.gguf](https://huggingface.co/afrideva/MiniChat-1.5-3B-GGUF/resolve/main/minichat-1.5-3b.q4_k_m.gguf) | q4_k_m | 1.85 GB | +| [minichat-1.5-3b.q5_k_m.gguf](https://huggingface.co/afrideva/MiniChat-1.5-3B-GGUF/resolve/main/minichat-1.5-3b.q5_k_m.gguf) | q5_k_m | 2.15 GB | +| [minichat-1.5-3b.q6_k.gguf](https://huggingface.co/afrideva/MiniChat-1.5-3B-GGUF/resolve/main/minichat-1.5-3b.q6_k.gguf) | q6_k | 2.48 GB | +| [minichat-1.5-3b.q8_0.gguf](https://huggingface.co/afrideva/MiniChat-1.5-3B-GGUF/resolve/main/minichat-1.5-3b.q8_0.gguf) | q8_0 | 3.21 GB | + + + +## Original Model Card: +## MiniChat-1.5-3B + +📑 [arXiv](https://arxiv.org/abs/2311.07052) | 👻 [GitHub](https://github.com/GeneZC/MiniMA) | 🤗 [HuggingFace-MiniMA](https://huggingface.co/GeneZC/MiniMA-3B) | 🤗 [HuggingFace-MiniChat](https://huggingface.co/GeneZC/MiniChat-3B) | 🤗 [HuggingFace-MiniChat-1.5](https://huggingface.co/GeneZC/MiniChat-1.5-3B) | 🤖 [ModelScope-MiniMA](https://modelscope.cn/models/GeneZC/MiniMA-3B) | 🤖 [ModelScope-MiniChat](https://modelscope.cn/models/GeneZC/MiniChat-3B) + +🆕 **Updates from MiniChat-3B**: +- better data mixture; +- use of [NEFTune](https://arxiv.org/abs/2310.05914); +- use of [DPO](https://arxiv.org/abs/2305.18290). + +❗ Must comply with LICENSE of LLaMA2 since it is derived from LLaMA2. + +A language model distilled and finetuned from an adapted version of LLaMA2-7B following "Towards the Law of Capacity Gap in Distilling Language Models". + +Outperforming a wide range of 3B competitors in GPT4 evaluation and even competing with several 7B chat models. + +teaser_b + +The following is an example code snippet to use MiniChat-3B: + +```python +import torch + +from transformers import AutoModelForCausalLM, AutoTokenizer + +from conversation import get_default_conv_template + +# MiniChat +tokenizer = AutoTokenizer.from_pretrained("GeneZC/MiniChat-3B", use_fast=False) +# GPU. +model = AutoModelForCausalLM.from_pretrained("GeneZC/MiniChat-3B", use_cache=True, device_map="auto", torch_dtype=torch.float16).eval() +# CPU. +# model = AutoModelForCausalLM.from_pretrained("GeneZC/MiniChat-3B", use_cache=True, device_map="cpu", torch_dtype=torch.float16).eval() + +conv = get_default_conv_template("minichat") + +question = "Implement a program to find the common elements in two arrays without using any extra data structures." +conv.append_message(conv.roles[0], question) +conv.append_message(conv.roles[1], None) +prompt = conv.get_prompt() +input_ids = tokenizer([prompt]).input_ids +output_ids = model.generate( + torch.as_tensor(input_ids).cuda(), + do_sample=True, + temperature=0.7, + max_new_tokens=1024, +) +output_ids = output_ids[0][len(input_ids[0]):] +output = tokenizer.decode(output_ids, skip_special_tokens=True).strip() +# output: "def common_elements(arr1, arr2):\n if len(arr1) == 0:\n return []\n if len(arr2) == 0:\n return arr1\n\n common_elements = []\n for element in arr1:\n if element in arr2:\n common_elements.append(element)\n\n return common_elements" +# Multiturn conversation could be realized by continuously appending questions to `conv`. +``` + +## Bibtex + +```bibtex +@article{zhang2023law, + title={Towards the Law of Capacity Gap in Distilling Language Models}, + author={Zhang, Chen and Song, Dawei and Ye, Zheyu and Gao, Yan}, + year={2023}, + url={https://arxiv.org/abs/2311.07052} +} +``` \ No newline at end of file diff --git a/minichat-1.5-3b.fp16.gguf b/minichat-1.5-3b.fp16.gguf new file mode 100644 index 0000000..67cf392 --- /dev/null +++ b/minichat-1.5-3b.fp16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3a50f741d0f11dbb58a370fee72beb13d3d312fc497d95b53b194a0373efca3 +size 6042292800 diff --git a/minichat-1.5-3b.q2_k.gguf b/minichat-1.5-3b.q2_k.gguf new file mode 100644 index 0000000..7734e99 --- /dev/null +++ b/minichat-1.5-3b.q2_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44fa484808b1ce7159b6b0955e8caa4622437dd26536b10aff8eb661b55cdff4 +size 1297187936 diff --git a/minichat-1.5-3b.q3_k_m.gguf b/minichat-1.5-3b.q3_k_m.gguf new file mode 100644 index 0000000..57b9c98 --- /dev/null +++ b/minichat-1.5-3b.q3_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f5767dc5022be419828ba59f252031421f4d78e9352f2a9c0529d914cb82c68 +size 1507578464 diff --git a/minichat-1.5-3b.q4_k_m.gguf b/minichat-1.5-3b.q4_k_m.gguf new file mode 100644 index 0000000..7c58604 --- /dev/null +++ b/minichat-1.5-3b.q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6720859e5b2d0dcef06c63fc7ee147841c4a0047728cbbd7555c6ae719247ee2 +size 1846655072 diff --git a/minichat-1.5-3b.q5_k_m.gguf b/minichat-1.5-3b.q5_k_m.gguf new file mode 100644 index 0000000..b3216cd --- /dev/null +++ b/minichat-1.5-3b.q5_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0157ae58c3e2abac76d842a8bfe0d842abc357b10e03d94d0d2165376f3ba1c9 +size 2153388128 diff --git a/minichat-1.5-3b.q6_k.gguf b/minichat-1.5-3b.q6_k.gguf new file mode 100644 index 0000000..13dcdbd --- /dev/null +++ b/minichat-1.5-3b.q6_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87f3c594128d13500d4a18e6b69c89f554910fbab4fefa4086e8555559933bcb +size 2479292000 diff --git a/minichat-1.5-3b.q8_0.gguf b/minichat-1.5-3b.q8_0.gguf new file mode 100644 index 0000000..39f13a5 --- /dev/null +++ b/minichat-1.5-3b.q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28c909c1f1c2f3affc5607439bc7219f4d23e4778d1a13c6bfe7d9d18d54df73 +size 3210768992