From f510c1e535892b7d81aea73b89e0d7a5b42dd8cf Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Wed, 22 Apr 2026 00:04:55 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: afrideva/llama-160m-GGUF Source: Original Platform --- .gitattributes | 42 +++++++++++++++++++++++++++++ README.md | 60 ++++++++++++++++++++++++++++++++++++++++++ llama-160m.fp16.gguf | 3 +++ llama-160m.q2_k.gguf | 3 +++ llama-160m.q3_k_m.gguf | 3 +++ llama-160m.q4_k_m.gguf | 3 +++ llama-160m.q5_k_m.gguf | 3 +++ llama-160m.q6_k.gguf | 3 +++ llama-160m.q8_0.gguf | 3 +++ 9 files changed, 123 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 llama-160m.fp16.gguf create mode 100644 llama-160m.q2_k.gguf create mode 100644 llama-160m.q3_k_m.gguf create mode 100644 llama-160m.q4_k_m.gguf create mode 100644 llama-160m.q5_k_m.gguf create mode 100644 llama-160m.q6_k.gguf create mode 100644 llama-160m.q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..897512f --- /dev/null +++ b/.gitattributes @@ -0,0 +1,42 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +llama-160m.fp16.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m.q2_k.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m.q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m.q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m.q6_k.gguf filter=lfs diff=lfs merge=lfs -text +llama-160m.q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..d1d62bb --- /dev/null +++ b/README.md @@ -0,0 +1,60 @@ +--- +base_model: JackFram/llama-160m +datasets: +- wikipedia +inference: false +language: +- en +license: other +model_creator: JackFram +model_name: llama-160m +pipeline_tag: text-generation +quantized_by: afrideva +tags: +- gguf +- ggml +- quantized +- q2_k +- q3_k_m +- q4_k_m +- q5_k_m +- q6_k +- q8_0 +--- +# JackFram/llama-160m-GGUF + +Quantized GGUF model files for [llama-160m](https://huggingface.co/JackFram/llama-160m) from [JackFram](https://huggingface.co/JackFram) + + +| Name | Quant method | Size | +| ---- | ---- | ---- | +| [llama-160m.fp16.gguf](https://huggingface.co/afrideva/llama-160m-GGUF/resolve/main/llama-160m.fp16.gguf) | fp16 | 326.58 MB | +| [llama-160m.q2_k.gguf](https://huggingface.co/afrideva/llama-160m-GGUF/resolve/main/llama-160m.q2_k.gguf) | q2_k | 77.23 MB | +| [llama-160m.q3_k_m.gguf](https://huggingface.co/afrideva/llama-160m-GGUF/resolve/main/llama-160m.q3_k_m.gguf) | q3_k_m | 87.54 MB | +| [llama-160m.q4_k_m.gguf](https://huggingface.co/afrideva/llama-160m-GGUF/resolve/main/llama-160m.q4_k_m.gguf) | q4_k_m | 104.03 MB | +| [llama-160m.q5_k_m.gguf](https://huggingface.co/afrideva/llama-160m-GGUF/resolve/main/llama-160m.q5_k_m.gguf) | q5_k_m | 119.04 MB | +| [llama-160m.q6_k.gguf](https://huggingface.co/afrideva/llama-160m-GGUF/resolve/main/llama-160m.q6_k.gguf) | q6_k | 135.00 MB | +| [llama-160m.q8_0.gguf](https://huggingface.co/afrideva/llama-160m-GGUF/resolve/main/llama-160m.q8_0.gguf) | q8_0 | 174.33 MB | + + + +## Original Model Card: +## Model description +This is a LLaMA-like model with only 160M parameters trained on Wikipedia and part of the C4-en and C4-realnewslike datasets. + +No evaluation has been conducted yet, so use it with care. + +The model is mainly developed as a base Small Speculative Model in the [SpecInfer](https://arxiv.org/abs/2305.09781) paper. + +## Citation +To cite the model, please use +```bibtex +@misc{miao2023specinfer, + title={SpecInfer: Accelerating Generative LLM Serving with Speculative Inference and Token Tree Verification}, + author={Xupeng Miao and Gabriele Oliaro and Zhihao Zhang and Xinhao Cheng and Zeyu Wang and Rae Ying Yee Wong and Zhuoming Chen and Daiyaan Arfeen and Reyna Abhyankar and Zhihao Jia}, + year={2023}, + eprint={2305.09781}, + archivePrefix={arXiv}, + primaryClass={cs.CL} +} +``` \ No newline at end of file diff --git a/llama-160m.fp16.gguf b/llama-160m.fp16.gguf new file mode 100644 index 0000000..e20c8d9 --- /dev/null +++ b/llama-160m.fp16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fd24b8f26823a09f8bc1e72d55e69e31ad29fbbaae316baed339bd85c60d638 +size 326576768 diff --git a/llama-160m.q2_k.gguf b/llama-160m.q2_k.gguf new file mode 100644 index 0000000..50f38e7 --- /dev/null +++ b/llama-160m.q2_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd7b0291fa567cbd77b3b232df286219771fc3551a019ffa11e64d6a99479ed7 +size 77227168 diff --git a/llama-160m.q3_k_m.gguf b/llama-160m.q3_k_m.gguf new file mode 100644 index 0000000..d5d2d15 --- /dev/null +++ b/llama-160m.q3_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:609ecd7d5de43c9d94bfd25f782c0d399768468bb0c8e0af8791d99bc2ed86b6 +size 87538336 diff --git a/llama-160m.q4_k_m.gguf b/llama-160m.q4_k_m.gguf new file mode 100644 index 0000000..ec53978 --- /dev/null +++ b/llama-160m.q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d507ac9c5bc2c917f83eda34b8ea8226c14880ebc726d40a6495f443e983a9af +size 104027296 diff --git a/llama-160m.q5_k_m.gguf b/llama-160m.q5_k_m.gguf new file mode 100644 index 0000000..ace6ca8 --- /dev/null +++ b/llama-160m.q5_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3b529dfa0a7c2bc52f9e5923364755a1586d6040f0f2370900c32f5ac99dabe +size 119043232 diff --git a/llama-160m.q6_k.gguf b/llama-160m.q6_k.gguf new file mode 100644 index 0000000..1bcfa21 --- /dev/null +++ b/llama-160m.q6_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25942c15fda9ae9226f51b3d2b327ca1d1ff9e60e31e625599c2a9141631e9ef +size 134997664 diff --git a/llama-160m.q8_0.gguf b/llama-160m.q8_0.gguf new file mode 100644 index 0000000..588e95a --- /dev/null +++ b/llama-160m.q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1adf4e3eba74804aad81bf51ab98daf3dffd15b48912d0e2f3885b985c028df7 +size 174328480