From 047d6fcbbf966b810e4ae713bce877cd40af4705 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 14 Apr 2026 01:32:03 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: afrideva/smol_llama-220M-openhermes-GGUF Source: Original Platform --- .gitattributes | 42 ++++++++++ README.md | 102 +++++++++++++++++++++++++ smol_llama-220m-openhermes.fp16.gguf | 3 + smol_llama-220m-openhermes.q2_k.gguf | 3 + smol_llama-220m-openhermes.q3_k_m.gguf | 3 + smol_llama-220m-openhermes.q4_k_m.gguf | 3 + smol_llama-220m-openhermes.q5_k_m.gguf | 3 + smol_llama-220m-openhermes.q6_k.gguf | 3 + smol_llama-220m-openhermes.q8_0.gguf | 3 + 9 files changed, 165 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 smol_llama-220m-openhermes.fp16.gguf create mode 100644 smol_llama-220m-openhermes.q2_k.gguf create mode 100644 smol_llama-220m-openhermes.q3_k_m.gguf create mode 100644 smol_llama-220m-openhermes.q4_k_m.gguf create mode 100644 smol_llama-220m-openhermes.q5_k_m.gguf create mode 100644 smol_llama-220m-openhermes.q6_k.gguf create mode 100644 smol_llama-220m-openhermes.q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..f52304b --- /dev/null +++ b/.gitattributes @@ -0,0 +1,42 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-openhermes.fp16.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-openhermes.q2_k.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-openhermes.q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-openhermes.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-openhermes.q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-openhermes.q6_k.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-openhermes.q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..dca282d --- /dev/null +++ b/README.md @@ -0,0 +1,102 @@ +--- +base_model: BEE-spoke-data/smol_llama-220M-openhermes +datasets: +- teknium/openhermes +inference: false +license: apache-2.0 +model_creator: BEE-spoke-data +model_name: smol_llama-220M-openhermes +pipeline_tag: text-generation +quantized_by: afrideva +tags: +- gguf +- ggml +- quantized +- q2_k +- q3_k_m +- q4_k_m +- q5_k_m +- q6_k +- q8_0 +widget: +- example_title: burritos + text: "Below is an instruction that describes a task, paired with an input that + provides further context. Write a response that appropriately completes the request. + \ \n \n### Instruction: \n \nWrite an ode to Chipotle burritos. \n \n### + Response: \n" +--- +# BEE-spoke-data/smol_llama-220M-openhermes-GGUF + +Quantized GGUF model files for [smol_llama-220M-openhermes](https://huggingface.co/BEE-spoke-data/smol_llama-220M-openhermes) from [BEE-spoke-data](https://huggingface.co/BEE-spoke-data) + + +| Name | Quant method | Size | +| ---- | ---- | ---- | +| [smol_llama-220m-openhermes.fp16.gguf](https://huggingface.co/afrideva/smol_llama-220M-openhermes-GGUF/resolve/main/smol_llama-220m-openhermes.fp16.gguf) | fp16 | 436.50 MB | +| [smol_llama-220m-openhermes.q2_k.gguf](https://huggingface.co/afrideva/smol_llama-220M-openhermes-GGUF/resolve/main/smol_llama-220m-openhermes.q2_k.gguf) | q2_k | 94.43 MB | +| [smol_llama-220m-openhermes.q3_k_m.gguf](https://huggingface.co/afrideva/smol_llama-220M-openhermes-GGUF/resolve/main/smol_llama-220m-openhermes.q3_k_m.gguf) | q3_k_m | 114.65 MB | +| [smol_llama-220m-openhermes.q4_k_m.gguf](https://huggingface.co/afrideva/smol_llama-220M-openhermes-GGUF/resolve/main/smol_llama-220m-openhermes.q4_k_m.gguf) | q4_k_m | 137.58 MB | +| [smol_llama-220m-openhermes.q5_k_m.gguf](https://huggingface.co/afrideva/smol_llama-220M-openhermes-GGUF/resolve/main/smol_llama-220m-openhermes.q5_k_m.gguf) | q5_k_m | 157.91 MB | +| [smol_llama-220m-openhermes.q6_k.gguf](https://huggingface.co/afrideva/smol_llama-220M-openhermes-GGUF/resolve/main/smol_llama-220m-openhermes.q6_k.gguf) | q6_k | 179.52 MB | +| [smol_llama-220m-openhermes.q8_0.gguf](https://huggingface.co/afrideva/smol_llama-220M-openhermes-GGUF/resolve/main/smol_llama-220m-openhermes.q8_0.gguf) | q8_0 | 232.28 MB | + + + +## Original Model Card: +# BEE-spoke-data/smol_llama-220M-openhermes + +> Please note that this is an experiment, and the model has limitations because it is smol. + + +prompt format is alpaca + + +``` +Below is an instruction that describes a task, paired with an input that +provides further context. Write a response that appropriately completes +the request. + +### Instruction: + +How can I increase my meme production/output? Currently, I only create them in ancient babylonian which is time consuming. + +### Inputs: + +### Response: +``` + +It was trained on inputs so if you have inputs (like some text to ask a question about) then include it under `### Inputs:` + + +## Example + +Output on the text above ^. The inference API is set to sample with low temp so you should see (_at least slightly_) different generations each time. + + +![image/png](https://cdn-uploads.huggingface.co/production/uploads/60bccec062080d33f875cd0c/0nFP2jsBkritnryKmI8NV.png) + +Note that the inference API parameters used here are an initial educated guess, and may be updated over time: + +```yml +inference: + parameters: + do_sample: true + renormalize_logits: true + temperature: 0.25 + top_p: 0.95 + top_k: 50 + min_new_tokens: 2 + max_new_tokens: 96 + repetition_penalty: 1.03 + no_repeat_ngram_size: 5 + epsilon_cutoff: 0.0008 +``` + +Feel free to experiment with the parameters using the model in Python and let us know if you have improved results with other params! + +## Data + +Note that **this checkpoint** was fine-tuned on `teknium/openhermes`, which is generated/synthetic data by an OpenAI model. This means usage of this checkpoint should follow their terms of use: https://openai.com/policies/terms-of-use + + +--- \ No newline at end of file diff --git a/smol_llama-220m-openhermes.fp16.gguf b/smol_llama-220m-openhermes.fp16.gguf new file mode 100644 index 0000000..0bbd199 --- /dev/null +++ b/smol_llama-220m-openhermes.fp16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9c186c2c573bb0d81ed87e17aee2a85b407223ea027ea29e1a3d2426472dc8f +size 436502176 diff --git a/smol_llama-220m-openhermes.q2_k.gguf b/smol_llama-220m-openhermes.q2_k.gguf new file mode 100644 index 0000000..0605908 --- /dev/null +++ b/smol_llama-220m-openhermes.q2_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c79d47dc4d7e7c4146f3baf1ac0dd01bd5fdcd82380f49ed47036ff4ad064f3 +size 94429888 diff --git a/smol_llama-220m-openhermes.q3_k_m.gguf b/smol_llama-220m-openhermes.q3_k_m.gguf new file mode 100644 index 0000000..5c2bf16 --- /dev/null +++ b/smol_llama-220m-openhermes.q3_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ccb08c83a0f1840923fd7aaa8bc037f8e7a046087ee5e168827bde5b0e253eb +size 114650816 diff --git a/smol_llama-220m-openhermes.q4_k_m.gguf b/smol_llama-220m-openhermes.q4_k_m.gguf new file mode 100644 index 0000000..308baa2 --- /dev/null +++ b/smol_llama-220m-openhermes.q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a804b9f64aabb28fb188683832a7773341a7ca2e2f94450850527c8cd13cb9f5 +size 137581248 diff --git a/smol_llama-220m-openhermes.q5_k_m.gguf b/smol_llama-220m-openhermes.q5_k_m.gguf new file mode 100644 index 0000000..c6661d8 --- /dev/null +++ b/smol_llama-220m-openhermes.q5_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:094269e1a081eec5d9dfa94b205b4f0e0ffd58052c43fb32a15083bbe796363d +size 157913792 diff --git a/smol_llama-220m-openhermes.q6_k.gguf b/smol_llama-220m-openhermes.q6_k.gguf new file mode 100644 index 0000000..a72e6d5 --- /dev/null +++ b/smol_llama-220m-openhermes.q6_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:844058b8eccbec057f62f26d3f525dd1fdb46f1be01163ee8eb8c0b8b1550b15 +size 179517120 diff --git a/smol_llama-220m-openhermes.q8_0.gguf b/smol_llama-220m-openhermes.q8_0.gguf new file mode 100644 index 0000000..c7b9484 --- /dev/null +++ b/smol_llama-220m-openhermes.q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6942bc7b12350d2c792fba01f300948226b15c7d993b6e677f589e286b450a4b +size 232275648