From 31c57c1a431d3038bb75ae105b784801eb85e8e3 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 28 Apr 2026 15:42:58 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: afrideva/Llama-160M-Chat-v1-GGUF Source: Original Platform --- .gitattributes | 42 +++++++++ README.md | 160 +++++++++++++++++++++++++++++++++ llama-160m-chat-v1.fp16.gguf | 3 + llama-160m-chat-v1.q2_k.gguf | 3 + llama-160m-chat-v1.q3_k_m.gguf | 3 + llama-160m-chat-v1.q4_k_m.gguf | 3 + llama-160m-chat-v1.q5_k_m.gguf | 3 + llama-160m-chat-v1.q6_k.gguf | 3 + llama-160m-chat-v1.q8_0.gguf | 3 + 9 files changed, 223 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 llama-160m-chat-v1.fp16.gguf create mode 100644 llama-160m-chat-v1.q2_k.gguf create mode 100644 llama-160m-chat-v1.q3_k_m.gguf create mode 100644 llama-160m-chat-v1.q4_k_m.gguf create mode 100644 llama-160m-chat-v1.q5_k_m.gguf create mode 100644 llama-160m-chat-v1.q6_k.gguf create mode 100644 llama-160m-chat-v1.q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..62d36f7 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,42 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +llama-160m-chat-v1.fp16.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m-chat-v1.q2_k.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m-chat-v1.q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m-chat-v1.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m-chat-v1.q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m-chat-v1.q6_k.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m-chat-v1.q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..ed4a4c7 --- /dev/null +++ b/README.md @@ -0,0 +1,160 @@ +--- +base_model: Felladrin/Llama-160M-Chat-v1 +datasets: +- ehartford/wizard_vicuna_70k_unfiltered +- totally-not-an-llm/EverythingLM-data-V3 +- Open-Orca/SlimOrca-Dedup +- databricks/databricks-dolly-15k +- THUDM/webglm-qa +inference: false +license: other +model_creator: Felladrin +model_name: Llama-160M-Chat-v1 +pipeline_tag: text-generation +quantized_by: afrideva +tags: +- text-generation +- gguf +- ggml +- quantized +- q2_k +- q3_k_m +- q4_k_m +- q5_k_m +- q6_k +- q8_0 +widget: +- text: "<|im_start|>system\nYou are a helpful assistant, who answers with empathy.<|im_end|>\n<|im_start|>user\nGot + a question for you!<|im_end|>\n<|im_start|>assistant\nSure! What's it?<|im_end|>\n<|im_start|>user\nWhy + do you love cats so much!? \U0001F408<|im_end|>\n<|im_start|>assistant" +- text: '<|im_start|>system + + You are a helpful assistant who answers user''s questions with empathy.<|im_end|> + + <|im_start|>user + + Who is Mona Lisa?<|im_end|> + + <|im_start|>assistant' +- text: '<|im_start|>system + + You are a helpful assistant who provides concise responses.<|im_end|> + + <|im_start|>user + + Heya!<|im_end|> + + <|im_start|>assistant + + Hi! How may I help you today?<|im_end|> + + <|im_start|>user + + I need to build a simple website. Where should I start learning about web development?<|im_end|> + + <|im_start|>assistant' +- text: '<|im_start|>user + + Invited some friends to come home today. Give me some ideas for games to play + with them!<|im_end|> + + <|im_start|>assistant' +- text: '<|im_start|>system + + You are a helpful assistant who answers user''s questions with details and curiosity.<|im_end|> + + <|im_start|>user + + What are some potential applications for quantum computing?<|im_end|> + + <|im_start|>assistant' +- text: '<|im_start|>system + + You are a helpful assistant who gives creative responses.<|im_end|> + + <|im_start|>user + + Write the specs of a game about mages in a fantasy world.<|im_end|> + + <|im_start|>assistant' +- text: '<|im_start|>system + + You are a helpful assistant who answers user''s questions with details.<|im_end|> + + <|im_start|>user + + Tell me about the pros and cons of social media.<|im_end|> + + <|im_start|>assistant' +- text: '<|im_start|>system + + You are a helpful assistant who answers user''s questions with confidence.<|im_end|> + + <|im_start|>user + + What is a dog?<|im_end|> + + <|im_start|>assistant + + A dog is a four-legged, domesticated animal that is a member of the class Mammalia, + which includes all mammals. Dogs are known for their loyalty, playfulness, and + ability to be trained for various tasks. They are also used for hunting, herding, + and as service animals.<|im_end|> + + <|im_start|>user + + What is the color of an apple?<|im_end|> + + <|im_start|>assistant' +--- +# Felladrin/Llama-160M-Chat-v1-GGUF + +Quantized GGUF model files for [Llama-160M-Chat-v1](https://huggingface.co/Felladrin/Llama-160M-Chat-v1) from [Felladrin](https://huggingface.co/Felladrin) + + +| Name | Quant method | Size | +| ---- | ---- | ---- | +| [llama-160m-chat-v1.fp16.gguf](https://huggingface.co/afrideva/Llama-160M-Chat-v1-GGUF/resolve/main/llama-160m-chat-v1.fp16.gguf) | fp16 | 326.58 MB | +| [llama-160m-chat-v1.q2_k.gguf](https://huggingface.co/afrideva/Llama-160M-Chat-v1-GGUF/resolve/main/llama-160m-chat-v1.q2_k.gguf) | q2_k | 77.23 MB | +| [llama-160m-chat-v1.q3_k_m.gguf](https://huggingface.co/afrideva/Llama-160M-Chat-v1-GGUF/resolve/main/llama-160m-chat-v1.q3_k_m.gguf) | q3_k_m | 87.54 MB | +| [llama-160m-chat-v1.q4_k_m.gguf](https://huggingface.co/afrideva/Llama-160M-Chat-v1-GGUF/resolve/main/llama-160m-chat-v1.q4_k_m.gguf) | q4_k_m | 104.03 MB | +| [llama-160m-chat-v1.q5_k_m.gguf](https://huggingface.co/afrideva/Llama-160M-Chat-v1-GGUF/resolve/main/llama-160m-chat-v1.q5_k_m.gguf) | q5_k_m | 119.04 MB | +| [llama-160m-chat-v1.q6_k.gguf](https://huggingface.co/afrideva/Llama-160M-Chat-v1-GGUF/resolve/main/llama-160m-chat-v1.q6_k.gguf) | q6_k | 135.00 MB | +| [llama-160m-chat-v1.q8_0.gguf](https://huggingface.co/afrideva/Llama-160M-Chat-v1-GGUF/resolve/main/llama-160m-chat-v1.q8_0.gguf) | q8_0 | 174.33 MB | + + + +## Original Model Card: +# A Llama Chat Model of 160M Parameters + +- Base model: [JackFram/llama-160m](https://huggingface.co/JackFram/llama-160m) +- Datasets: + - [ehartford/wizard_vicuna_70k_unfiltered](https://huggingface.co/datasets/ehartford/wizard_vicuna_70k_unfiltered) + - [totally-not-an-llm/EverythingLM-data-V3](https://huggingface.co/datasets/totally-not-an-llm/EverythingLM-data-V3) + - [Open-Orca/SlimOrca-Dedup](https://huggingface.co/datasets/Open-Orca/SlimOrca-Dedup) + - [databricks/databricks-dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k) + - [THUDM/webglm-qa](https://huggingface.co/datasets/THUDM/webglm-qa) +- Availability in other ML formats: + - GGUF: [Felladrin/gguf-Llama-160M-Chat-v1](https://huggingface.co/Felladrin/gguf-Llama-160M-Chat-v1) + - ONNX: [Felladrin/onnx-Llama-160M-Chat-v1](https://huggingface.co/Felladrin/onnx-Llama-160M-Chat-v1) + +## Recommended Prompt Format + +The recommended prompt format is as follows: + +``` +<|im_start|>system +{system_message}<|im_end|> +<|im_start|>user +{user_message}<|im_end|> +<|im_start|>assistant +``` + +## Recommended Inference Parameters + +To get the best results, prefer using [contrastive search](https://huggingface.co/docs/transformers/main/en/generation_strategies#contrastive-search) for inference: + +```yml +penalty_alpha: 0.5 +top_k: 5 +``` \ No newline at end of file diff --git a/llama-160m-chat-v1.fp16.gguf b/llama-160m-chat-v1.fp16.gguf new file mode 100644 index 0000000..8e18362 --- /dev/null +++ b/llama-160m-chat-v1.fp16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bbd042cdffc7ed19530edee18dc7705878392d6614bdcc405f958b96527120b2 +size 326576768 diff --git a/llama-160m-chat-v1.q2_k.gguf b/llama-160m-chat-v1.q2_k.gguf new file mode 100644 index 0000000..7e54701 --- /dev/null +++ b/llama-160m-chat-v1.q2_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae7c34f51e1689f7f379ab510e37c86a2ef624c5033febb5e6f2e7afb9ddd074 +size 77227168 diff --git a/llama-160m-chat-v1.q3_k_m.gguf b/llama-160m-chat-v1.q3_k_m.gguf new file mode 100644 index 0000000..6073809 --- /dev/null +++ b/llama-160m-chat-v1.q3_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b8b6507b8dcf0f01a68eac05386956a7e7a8a2784f6ea1efecdbea9b27069fe +size 87538336 diff --git a/llama-160m-chat-v1.q4_k_m.gguf b/llama-160m-chat-v1.q4_k_m.gguf new file mode 100644 index 0000000..0c26b9a --- /dev/null +++ b/llama-160m-chat-v1.q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:718824e01a966a6a0827eead47670cc91488019e7c2ee4cecc46d125483ebbb9 +size 104027296 diff --git a/llama-160m-chat-v1.q5_k_m.gguf b/llama-160m-chat-v1.q5_k_m.gguf new file mode 100644 index 0000000..97babf3 --- /dev/null +++ b/llama-160m-chat-v1.q5_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04ce0e7e2d305e0f8131047ef34385917eb2e9b080f56266459aa72f6b5e52b2 +size 119043232 diff --git a/llama-160m-chat-v1.q6_k.gguf b/llama-160m-chat-v1.q6_k.gguf new file mode 100644 index 0000000..2bd4fe3 --- /dev/null +++ b/llama-160m-chat-v1.q6_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d54692b5590f808c47a411e32c98747b89a863515142e322841b58e04814d3b9 +size 134997664 diff --git a/llama-160m-chat-v1.q8_0.gguf b/llama-160m-chat-v1.q8_0.gguf new file mode 100644 index 0000000..7c5da53 --- /dev/null +++ b/llama-160m-chat-v1.q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d7c6e22036397f6f9a5be725bdeb6fb6baf6a20c95212c15fb21ddcc0cca505 +size 174328480