commit ffb70ffd87311eb67767560cd654d8f41b92e62d Author: ModelHub XC Date: Wed Apr 22 00:04:55 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: afrideva/Smol-Llama-101M-Chat-v1-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..f499438 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,42 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +smol-llama-101m-chat-v1.fp16.gguf filter=lfs diff=lfs merge=lfs -text +smol-llama-101m-chat-v1.q2_k.gguf filter=lfs diff=lfs merge=lfs -text +smol-llama-101m-chat-v1.q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol-llama-101m-chat-v1.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol-llama-101m-chat-v1.q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol-llama-101m-chat-v1.q6_k.gguf filter=lfs diff=lfs merge=lfs -text +smol-llama-101m-chat-v1.q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..b78782a --- /dev/null +++ b/README.md @@ -0,0 +1,188 @@ +--- +base_model: Felladrin/Smol-Llama-101M-Chat-v1 +datasets: +- Open-Orca/SlimOrca-Dedup +- VMware/open-instruct +- LDJnr/Capybara +- cognitivecomputations/ultrachat-uncensored +- starfishmedical/webGPT_x_dolly +- THUDM/webglm-qa +inference: false +language: +- en +license: apache-2.0 +model_creator: Felladrin +model_name: Smol-Llama-101M-Chat-v1 +pipeline_tag: text-generation +quantized_by: afrideva +tags: +- text-generation +- gguf +- ggml +- quantized +- q2_k +- q3_k_m +- q4_k_m +- q5_k_m +- q6_k +- q8_0 +widget: +- text: '<|im_start|>system + + You are a helpful assistant who gives creative responses.<|im_end|> + + <|im_start|>user + + Write the background story of a game about wizards and llamas in a sci-fi world.<|im_end|> + + <|im_start|>assistant' +- text: '<|im_start|>system + + A friendly chat between a user and an assistant.<|im_end|> + + <|im_start|>user + + Got a question for you!<|im_end|> + + <|im_start|>assistant + + Sure! What''s it?<|im_end|> + + <|im_start|>user + + I need to build a simple website. Where should I start learning about web development?<|im_end|> + + <|im_start|>assistant' +- text: '<|im_start|>system + + You are a helpful assistant who provides concise answers to the user''s questions.<|im_end|> + + <|im_start|>user + + How to become more healthy?<|im_end|> + + <|im_start|>assistant' +- text: '<|im_start|>system + + You are a helpful assistant, who always answers with empathy.<|im_end|> + + <|im_start|>user + + List the pros and cons of social media.<|im_end|> + + <|im_start|>assistant' +- text: '<|im_start|>system + + You are a helpful assistant, who always answers with empathy.<|im_end|> + + <|im_start|>user + + Hello!<|im_end|> + + <|im_start|>assistant + + Hi! How can I help you today?<|im_end|> + + <|im_start|>user + + Take a look at the info below. + + + - The tape inside the VHS cassettes is very delicate and can be easily ruined, + making them unplayable and unrepairable. The reason the tape deteriorates is that + the magnetic charge needed for them to work is not permanent, and the magnetic + particles end up losing their charge in a process known as remanence decay. These + particles could also become demagnetised via being stored too close to a magnetic + source. + + - One of the most significant issues with VHS tapes is that they have moving parts, + meaning that there are more occasions when something can go wrong, damaging your + footage or preventing it from playing back. The tape itself is a prominent cause + of this, and tape slippage can occur. Tapes slippage can be caused when the tape + loses its tension, or it has become warped. These problems can occur in storage + due to high temperatures or frequent changes in humidity. + + - VHS tapes deteriorate over time from infrequent or overuse. Neglect means mold + and dirt, while overuse can lead to scratches and technical difficulties. This + is why old VHS tapes inevitably experience malfunctions after a long period of + time. Usually anywhere between 10 to 25+ years. + + - Some VHS tapes like newer mini DVs and Digital 8 tapes can suffer from digital + corruption, meaning that the footage becomes lost and cannot be recovered. These + tapes were the steppingstone from VHS to the digital age when capturing footage + straight to digital became the norm. Unfortunately,they are susceptible to digital + corruption, which causes video pixilation and/or loss of audio.<|im_end|> + + <|im_start|>assistant + + Alright!<|im_end|> + + <|im_start|>user + + Now I''m going to write my question, and if the info above is useful, you can + use them in your response. + + Ready?<|im_end|> + + <|im_start|>assistant + + Ready for your question!<|im_end|> + + <|im_start|>user + + Why do VHS tapes deteriorate over time?<|im_end|> + + <|im_start|>assistant' +--- +# Felladrin/Smol-Llama-101M-Chat-v1-GGUF + +Quantized GGUF model files for [Smol-Llama-101M-Chat-v1](https://huggingface.co/Felladrin/Smol-Llama-101M-Chat-v1) from [Felladrin](https://huggingface.co/Felladrin) + + +| Name | Quant method | Size | +| ---- | ---- | ---- | +| [smol-llama-101m-chat-v1.fp16.gguf](https://huggingface.co/afrideva/Smol-Llama-101M-Chat-v1-GGUF/resolve/main/smol-llama-101m-chat-v1.fp16.gguf) | fp16 | 204.25 MB | +| [smol-llama-101m-chat-v1.q2_k.gguf](https://huggingface.co/afrideva/Smol-Llama-101M-Chat-v1-GGUF/resolve/main/smol-llama-101m-chat-v1.q2_k.gguf) | q2_k | 51.90 MB | +| [smol-llama-101m-chat-v1.q3_k_m.gguf](https://huggingface.co/afrideva/Smol-Llama-101M-Chat-v1-GGUF/resolve/main/smol-llama-101m-chat-v1.q3_k_m.gguf) | q3_k_m | 58.04 MB | +| [smol-llama-101m-chat-v1.q4_k_m.gguf](https://huggingface.co/afrideva/Smol-Llama-101M-Chat-v1-GGUF/resolve/main/smol-llama-101m-chat-v1.q4_k_m.gguf) | q4_k_m | 66.38 MB | +| [smol-llama-101m-chat-v1.q5_k_m.gguf](https://huggingface.co/afrideva/Smol-Llama-101M-Chat-v1-GGUF/resolve/main/smol-llama-101m-chat-v1.q5_k_m.gguf) | q5_k_m | 75.31 MB | +| [smol-llama-101m-chat-v1.q6_k.gguf](https://huggingface.co/afrideva/Smol-Llama-101M-Chat-v1-GGUF/resolve/main/smol-llama-101m-chat-v1.q6_k.gguf) | q6_k | 84.80 MB | +| [smol-llama-101m-chat-v1.q8_0.gguf](https://huggingface.co/afrideva/Smol-Llama-101M-Chat-v1-GGUF/resolve/main/smol-llama-101m-chat-v1.q8_0.gguf) | q8_0 | 109.33 MB | + + + +## Original Model Card: +# A Llama Chat Model of 101M Parameters + +- Base model: [BEE-spoke-data/smol_llama-101M-GQA](https://huggingface.co/BEE-spoke-data/smol_llama-101M-GQA) +- Datasets: + - [Open-Orca/SlimOrca-Dedup](https://huggingface.co/datasets/Open-Orca/SlimOrca-Dedup) + - [VMware/open-instruct](https://huggingface.co/datasets/VMware/open-instruct) + - [LDJnr/Capybara](https://huggingface.co/datasets/LDJnr/Capybara) + - [cognitivecomputations/ultrachat-uncensored](https://huggingface.co/datasets/cognitivecomputations/ultrachat-uncensored) + - [starfishmedical/webGPT_x_dolly](https://huggingface.co/datasets/starfishmedical/webGPT_x_dolly) + - [THUDM/webglm-qa](https://huggingface.co/datasets/THUDM/webglm-qa) +- Availability in other ML formats: + - ONNX: [Felladrin/onnx-Smol-Llama-101M-Chat-v1](https://huggingface.co/Felladrin/onnx-Smol-Llama-101M-Chat-v1) + +## Recommended Prompt Format + +The recommended prompt format is as follows: + +``` +<|im_start|>system +{system_message}<|im_end|> +<|im_start|>user +{user_message}<|im_end|> +<|im_start|>assistant +``` + +## Recommended Inference Parameters + +To get the best results, add special tokens and prefer using [contrastive search](https://huggingface.co/docs/transformers/main/en/generation_strategies#contrastive-search) for inference: + +```yml +add_special_tokens: true +penalty_alpha: 0.5 +top_k: 5 +``` \ No newline at end of file diff --git a/smol-llama-101m-chat-v1.fp16.gguf b/smol-llama-101m-chat-v1.fp16.gguf new file mode 100644 index 0000000..2de4598 --- /dev/null +++ b/smol-llama-101m-chat-v1.fp16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34132a21850a13f16dc1a766e01bb20d5c257c06c66b45476b3ff7707363d220 +size 204250464 diff --git a/smol-llama-101m-chat-v1.q2_k.gguf b/smol-llama-101m-chat-v1.q2_k.gguf new file mode 100644 index 0000000..22aec32 --- /dev/null +++ b/smol-llama-101m-chat-v1.q2_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:718c042c8e3837ca9e75084058bbbc9a4986a31bbed5824fccebbab32322d9f3 +size 51904640 diff --git a/smol-llama-101m-chat-v1.q3_k_m.gguf b/smol-llama-101m-chat-v1.q3_k_m.gguf new file mode 100644 index 0000000..5a5c991 --- /dev/null +++ b/smol-llama-101m-chat-v1.q3_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ccc87c355bb639ecf2821b4a2060ccbc7071db6a8e34d56ca26323aeb0e29fa5 +size 58035584 diff --git a/smol-llama-101m-chat-v1.q4_k_m.gguf b/smol-llama-101m-chat-v1.q4_k_m.gguf new file mode 100644 index 0000000..bc30be6 --- /dev/null +++ b/smol-llama-101m-chat-v1.q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68f70b3f9aa690c40c412dea5c03a22f971cde81bca3eef588dd5a1d0412e64c +size 66378368 diff --git a/smol-llama-101m-chat-v1.q5_k_m.gguf b/smol-llama-101m-chat-v1.q5_k_m.gguf new file mode 100644 index 0000000..86368af --- /dev/null +++ b/smol-llama-101m-chat-v1.q5_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0102b97cdee1a411bfa8c5a28cd8bf3f7d6d32c7a02d808061ed2b762e7fa0f7 +size 75311744 diff --git a/smol-llama-101m-chat-v1.q6_k.gguf b/smol-llama-101m-chat-v1.q6_k.gguf new file mode 100644 index 0000000..f3c451f --- /dev/null +++ b/smol-llama-101m-chat-v1.q6_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a3be3781da878e36417b7d2789083855a09029a4a2db04856e2494669f4b1ea +size 84803456 diff --git a/smol-llama-101m-chat-v1.q8_0.gguf b/smol-llama-101m-chat-v1.q8_0.gguf new file mode 100644 index 0000000..e926031 --- /dev/null +++ b/smol-llama-101m-chat-v1.q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a9d4ea4169d365287e42f54e64c6142efa2ee31bf8b96dba1b880b33b27cdbb +size 109325696