From 1c0f5553d4caba77761004748dbd6bf8ef2c23b6 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 28 Apr 2026 23:10:31 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: afrideva/smol_llama-220M-open_instruct-GGUF Source: Original Platform --- .gitattributes | 42 ++++++++++ README.md | 99 +++++++++++++++++++++++ smol_llama-220m-open_instruct.fp16.gguf | 3 + smol_llama-220m-open_instruct.q2_k.gguf | 3 + smol_llama-220m-open_instruct.q3_k_m.gguf | 3 + smol_llama-220m-open_instruct.q4_k_m.gguf | 3 + smol_llama-220m-open_instruct.q5_k_m.gguf | 3 + smol_llama-220m-open_instruct.q6_k.gguf | 3 + smol_llama-220m-open_instruct.q8_0.gguf | 3 + 9 files changed, 162 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 smol_llama-220m-open_instruct.fp16.gguf create mode 100644 smol_llama-220m-open_instruct.q2_k.gguf create mode 100644 smol_llama-220m-open_instruct.q3_k_m.gguf create mode 100644 smol_llama-220m-open_instruct.q4_k_m.gguf create mode 100644 smol_llama-220m-open_instruct.q5_k_m.gguf create mode 100644 smol_llama-220m-open_instruct.q6_k.gguf create mode 100644 smol_llama-220m-open_instruct.q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..1b181a9 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,42 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-open_instruct.fp16.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-open_instruct.q2_k.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-open_instruct.q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-open_instruct.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-open_instruct.q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-open_instruct.q6_k.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-open_instruct.q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..7e06295 --- /dev/null +++ b/README.md @@ -0,0 +1,99 @@ +--- +base_model: BEE-spoke-data/smol_llama-220M-open_instruct +datasets: +- VMware/open-instruct +inference: false +license: apache-2.0 +model_creator: BEE-spoke-data +model_name: smol_llama-220M-open_instruct +pipeline_tag: text-generation +quantized_by: afrideva +tags: +- gguf +- ggml +- quantized +- q2_k +- q3_k_m +- q4_k_m +- q5_k_m +- q6_k +- q8_0 +widget: +- example_title: burritos + text: "Below is an instruction that describes a task, paired with an input that + provides further context. Write a response that appropriately completes the request. + \ \n \n### Instruction: \n \nWrite an ode to Chipotle burritos. \n \n### + Response: \n" +--- +# BEE-spoke-data/smol_llama-220M-open_instruct-GGUF + +Quantized GGUF model files for [smol_llama-220M-open_instruct](https://huggingface.co/BEE-spoke-data/smol_llama-220M-open_instruct) from [BEE-spoke-data](https://huggingface.co/BEE-spoke-data) + + +| Name | Quant method | Size | +| ---- | ---- | ---- | +| [smol_llama-220m-open_instruct.fp16.gguf](https://huggingface.co/afrideva/smol_llama-220M-open_instruct-GGUF/resolve/main/smol_llama-220m-open_instruct.fp16.gguf) | fp16 | 436.50 MB | +| [smol_llama-220m-open_instruct.q2_k.gguf](https://huggingface.co/afrideva/smol_llama-220M-open_instruct-GGUF/resolve/main/smol_llama-220m-open_instruct.q2_k.gguf) | q2_k | 94.43 MB | +| [smol_llama-220m-open_instruct.q3_k_m.gguf](https://huggingface.co/afrideva/smol_llama-220M-open_instruct-GGUF/resolve/main/smol_llama-220m-open_instruct.q3_k_m.gguf) | q3_k_m | 114.65 MB | +| [smol_llama-220m-open_instruct.q4_k_m.gguf](https://huggingface.co/afrideva/smol_llama-220M-open_instruct-GGUF/resolve/main/smol_llama-220m-open_instruct.q4_k_m.gguf) | q4_k_m | 137.58 MB | +| [smol_llama-220m-open_instruct.q5_k_m.gguf](https://huggingface.co/afrideva/smol_llama-220M-open_instruct-GGUF/resolve/main/smol_llama-220m-open_instruct.q5_k_m.gguf) | q5_k_m | 157.91 MB | +| [smol_llama-220m-open_instruct.q6_k.gguf](https://huggingface.co/afrideva/smol_llama-220M-open_instruct-GGUF/resolve/main/smol_llama-220m-open_instruct.q6_k.gguf) | q6_k | 179.52 MB | +| [smol_llama-220m-open_instruct.q8_0.gguf](https://huggingface.co/afrideva/smol_llama-220M-open_instruct-GGUF/resolve/main/smol_llama-220m-open_instruct.q8_0.gguf) | q8_0 | 232.28 MB | + + + +## Original Model Card: +# BEE-spoke-data/smol_llama-220M-open_instruct + +> Please note that this is an experiment, and the model has limitations because it is smol. + + +prompt format is alpaca. + + +``` +Below is an instruction that describes a task, paired with an input that +provides further context. Write a response that appropriately completes +the request. + +### Instruction: + +How can I increase my meme production/output? Currently, I only create them in ancient babylonian which is time consuming. + +### Response: +``` + +This was **not** trained using a separate 'inputs' field (as `VMware/open-instruct` doesn't use one). + + +## Example + +Output on the text above ^. The inference API is set to sample with low temp so you should see (_at least slightly_) different generations each time. + + +![image/png](https://cdn-uploads.huggingface.co/production/uploads/60bccec062080d33f875cd0c/MdOB7TD5UosPGZvdZWG0I.png) + +Note that the inference API parameters used here are an initial educated guess, and may be updated over time: + +```yml +inference: + parameters: + do_sample: true + renormalize_logits: true + temperature: 0.25 + top_p: 0.95 + top_k: 50 + min_new_tokens: 2 + max_new_tokens: 96 + repetition_penalty: 1.04 + no_repeat_ngram_size: 6 + epsilon_cutoff: 0.0006 +``` + +Feel free to experiment with the parameters using the model in Python and let us know if you have improved results with other params! + +## Data + +This was trained on `VMware/open-instruct` so do whatever you want, provided it falls under the base apache-2.0 license :) + +--- \ No newline at end of file diff --git a/smol_llama-220m-open_instruct.fp16.gguf b/smol_llama-220m-open_instruct.fp16.gguf new file mode 100644 index 0000000..ecc2801 --- /dev/null +++ b/smol_llama-220m-open_instruct.fp16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31883f6c6f6ef3872ba3208cac116d87dab706d0769915e085df29a138dcc93e +size 436502176 diff --git a/smol_llama-220m-open_instruct.q2_k.gguf b/smol_llama-220m-open_instruct.q2_k.gguf new file mode 100644 index 0000000..62eb9a9 --- /dev/null +++ b/smol_llama-220m-open_instruct.q2_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b9c2c2ddfb150c583abf5f06127f4fa749fa2a1547e0d77d7e2aa358c657a29 +size 94429888 diff --git a/smol_llama-220m-open_instruct.q3_k_m.gguf b/smol_llama-220m-open_instruct.q3_k_m.gguf new file mode 100644 index 0000000..3720b5f --- /dev/null +++ b/smol_llama-220m-open_instruct.q3_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e271fd56c134f77d7065a93790319fa5290c934d754c9bdadd228a4b58d0569 +size 114650816 diff --git a/smol_llama-220m-open_instruct.q4_k_m.gguf b/smol_llama-220m-open_instruct.q4_k_m.gguf new file mode 100644 index 0000000..1c39ae0 --- /dev/null +++ b/smol_llama-220m-open_instruct.q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0cb254c8c99f32edca193aaa1b8c6449e28d60ef2bc5bbac5e1ec1eab65480db +size 137581248 diff --git a/smol_llama-220m-open_instruct.q5_k_m.gguf b/smol_llama-220m-open_instruct.q5_k_m.gguf new file mode 100644 index 0000000..c441ce3 --- /dev/null +++ b/smol_llama-220m-open_instruct.q5_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74d2ba6df51b5bd802766e6a72eb5205cfb4e005fa56ebd8fe7810b211204ef1 +size 157913792 diff --git a/smol_llama-220m-open_instruct.q6_k.gguf b/smol_llama-220m-open_instruct.q6_k.gguf new file mode 100644 index 0000000..e48db43 --- /dev/null +++ b/smol_llama-220m-open_instruct.q6_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca8257a5e30be964a8e90b370ebc979b6750da63262046b500ef1970a9d30145 +size 179517120 diff --git a/smol_llama-220m-open_instruct.q8_0.gguf b/smol_llama-220m-open_instruct.q8_0.gguf new file mode 100644 index 0000000..bf8398e --- /dev/null +++ b/smol_llama-220m-open_instruct.q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f00061aae08cc9e1981002744acee20dd8f1a9c89a7608040490fca0ebae1188 +size 232275648