From 173aa3926572a8fe0ef5eb38e84e554a6e778509 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 5 May 2026 09:50:04 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: afrideva/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF Source: Original Platform --- .gitattributes | 42 +++++++++++ README.md | 71 +++++++++++++++++++ ...-1.1b-intermediate-step-1431k-3t.fp16.gguf | 3 + ...-1.1b-intermediate-step-1431k-3t.q2_k.gguf | 3 + ....1b-intermediate-step-1431k-3t.q3_k_m.gguf | 3 + ....1b-intermediate-step-1431k-3t.q4_k_m.gguf | 3 + ....1b-intermediate-step-1431k-3t.q5_k_m.gguf | 3 + ...-1.1b-intermediate-step-1431k-3t.q6_k.gguf | 3 + ...-1.1b-intermediate-step-1431k-3t.q8_0.gguf | 3 + 9 files changed, 134 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 tinyllama-1.1b-intermediate-step-1431k-3t.fp16.gguf create mode 100644 tinyllama-1.1b-intermediate-step-1431k-3t.q2_k.gguf create mode 100644 tinyllama-1.1b-intermediate-step-1431k-3t.q3_k_m.gguf create mode 100644 tinyllama-1.1b-intermediate-step-1431k-3t.q4_k_m.gguf create mode 100644 tinyllama-1.1b-intermediate-step-1431k-3t.q5_k_m.gguf create mode 100644 tinyllama-1.1b-intermediate-step-1431k-3t.q6_k.gguf create mode 100644 tinyllama-1.1b-intermediate-step-1431k-3t.q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..24f8074 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,42 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +tinyllama-1.1b-intermediate-step-1431k-3t.fp16.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-1.1b-intermediate-step-1431k-3t.q2_k.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-1.1b-intermediate-step-1431k-3t.q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-1.1b-intermediate-step-1431k-3t.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-1.1b-intermediate-step-1431k-3t.q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-1.1b-intermediate-step-1431k-3t.q6_k.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-1.1b-intermediate-step-1431k-3t.q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..9f44149 --- /dev/null +++ b/README.md @@ -0,0 +1,71 @@ +--- +base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T +datasets: +- cerebras/SlimPajama-627B +- bigcode/starcoderdata +inference: false +language: +- en +license: apache-2.0 +model_creator: TinyLlama +model_name: TinyLlama-1.1B-intermediate-step-1431k-3T +pipeline_tag: text-generation +quantized_by: afrideva +tags: +- gguf +- ggml +- quantized +- q2_k +- q3_k_m +- q4_k_m +- q5_k_m +- q6_k +- q8_0 +--- +# TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF + +Quantized GGUF model files for [TinyLlama-1.1B-intermediate-step-1431k-3T](https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T) from [TinyLlama](https://huggingface.co/TinyLlama) + + +| Name | Quant method | Size | +| ---- | ---- | ---- | +| [tinyllama-1.1b-intermediate-step-1431k-3t.fp16.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF/resolve/main/tinyllama-1.1b-intermediate-step-1431k-3t.fp16.gguf) | fp16 | 2.20 GB | +| [tinyllama-1.1b-intermediate-step-1431k-3t.q2_k.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF/resolve/main/tinyllama-1.1b-intermediate-step-1431k-3t.q2_k.gguf) | q2_k | 483.12 MB | +| [tinyllama-1.1b-intermediate-step-1431k-3t.q3_k_m.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF/resolve/main/tinyllama-1.1b-intermediate-step-1431k-3t.q3_k_m.gguf) | q3_k_m | 550.82 MB | +| [tinyllama-1.1b-intermediate-step-1431k-3t.q4_k_m.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF/resolve/main/tinyllama-1.1b-intermediate-step-1431k-3t.q4_k_m.gguf) | q4_k_m | 668.79 MB | +| [tinyllama-1.1b-intermediate-step-1431k-3t.q5_k_m.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF/resolve/main/tinyllama-1.1b-intermediate-step-1431k-3t.q5_k_m.gguf) | q5_k_m | 783.02 MB | +| [tinyllama-1.1b-intermediate-step-1431k-3t.q6_k.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF/resolve/main/tinyllama-1.1b-intermediate-step-1431k-3t.q6_k.gguf) | q6_k | 904.39 MB | +| [tinyllama-1.1b-intermediate-step-1431k-3t.q8_0.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF/resolve/main/tinyllama-1.1b-intermediate-step-1431k-3t.q8_0.gguf) | q8_0 | 1.17 GB | + + + +## Original Model Card: +
+ +# TinyLlama-1.1B +
+ +https://github.com/jzhang38/TinyLlama + +The TinyLlama project aims to **pretrain** a **1.1B Llama model on 3 trillion tokens**. With some proper optimization, we can achieve this within a span of "just" 90 days using 16 A100-40G GPUs 🚀🚀. The training has started on 2023-09-01. + +
+ +
+ +We adopted exactly the same architecture and tokenizer as Llama 2. This means TinyLlama can be plugged and played in many open-source projects built upon Llama. Besides, TinyLlama is compact with only 1.1B parameters. This compactness allows it to cater to a multitude of applications demanding a restricted computation and memory footprint. + +#### This Collection +This collection contains all checkpoints after the 1T fix. Branch name indicates the step and number of tokens seen. + +#### Eval + +| Model | Pretrain Tokens | HellaSwag | Obqa | WinoGrande | ARC_c | ARC_e | boolq | piqa | avg | +|-------------------------------------------|-----------------|-----------|------|------------|-------|-------|-------|------|-----| +| Pythia-1.0B | 300B | 47.16 | 31.40| 53.43 | 27.05 | 48.99 | 60.83 | 69.21 | 48.30 | +| TinyLlama-1.1B-intermediate-step-50K-104b | 103B | 43.50 | 29.80| 53.28 | 24.32 | 44.91 | 59.66 | 67.30 | 46.11| +| TinyLlama-1.1B-intermediate-step-240k-503b| 503B | 49.56 |31.40 |55.80 |26.54 |48.32 |56.91 |69.42 | 48.28 | +| TinyLlama-1.1B-intermediate-step-480k-1007B | 1007B | 52.54 | 33.40 | 55.96 | 27.82 | 52.36 | 59.54 | 69.91 | 50.22 | +| TinyLlama-1.1B-intermediate-step-715k-1.5T | 1.5T | 53.68 | 35.20 | 58.33 | 29.18 | 51.89 | 59.08 | 71.65 | 51.29 | +| TinyLlama-1.1B-intermediate-step-955k-2T | 2T | 54.63 | 33.40 | 56.83 | 28.07 | 54.67 | 63.21 | 70.67 | 51.64 | +| **TinyLlama-1.1B-intermediate-step-1195k-token-2.5T** | **2.5T** | **58.96** | **34.40** | **58.72** | **31.91** | **56.78** | **63.21** | **73.07** | **53.86**| \ No newline at end of file diff --git a/tinyllama-1.1b-intermediate-step-1431k-3t.fp16.gguf b/tinyllama-1.1b-intermediate-step-1431k-3t.fp16.gguf new file mode 100644 index 0000000..cc988d8 --- /dev/null +++ b/tinyllama-1.1b-intermediate-step-1431k-3t.fp16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:445632c2a3176d8dde39d1ac42f6d22469ee771dfea7a5ce97ad01b44fa37f30 +size 2201990016 diff --git a/tinyllama-1.1b-intermediate-step-1431k-3t.q2_k.gguf b/tinyllama-1.1b-intermediate-step-1431k-3t.q2_k.gguf new file mode 100644 index 0000000..baae238 --- /dev/null +++ b/tinyllama-1.1b-intermediate-step-1431k-3t.q2_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf8a430ca83b1ce6cb318d6ee6e31e487e9b7fc6c0060883ef68010a39c5b467 +size 483115936 diff --git a/tinyllama-1.1b-intermediate-step-1431k-3t.q3_k_m.gguf b/tinyllama-1.1b-intermediate-step-1431k-3t.q3_k_m.gguf new file mode 100644 index 0000000..77cf16c --- /dev/null +++ b/tinyllama-1.1b-intermediate-step-1431k-3t.q3_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb5d2229edb6f2d6dcadda860a6c8c416f556e36052486edc79262c174b59216 +size 550818720 diff --git a/tinyllama-1.1b-intermediate-step-1431k-3t.q4_k_m.gguf b/tinyllama-1.1b-intermediate-step-1431k-3t.q4_k_m.gguf new file mode 100644 index 0000000..1e950fd --- /dev/null +++ b/tinyllama-1.1b-intermediate-step-1431k-3t.q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:414c0dcb0f5798236226b1438997c6666da746a0fba03e323eaa77e95da57adc +size 668787616 diff --git a/tinyllama-1.1b-intermediate-step-1431k-3t.q5_k_m.gguf b/tinyllama-1.1b-intermediate-step-1431k-3t.q5_k_m.gguf new file mode 100644 index 0000000..a95ac99 --- /dev/null +++ b/tinyllama-1.1b-intermediate-step-1431k-3t.q5_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05436d7b35387d38450bbc6a01d5693249db6aa58c3e93e8414482571ab81941 +size 783016864 diff --git a/tinyllama-1.1b-intermediate-step-1431k-3t.q6_k.gguf b/tinyllama-1.1b-intermediate-step-1431k-3t.q6_k.gguf new file mode 100644 index 0000000..80cfd0a --- /dev/null +++ b/tinyllama-1.1b-intermediate-step-1431k-3t.q6_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e646cba128b79ac34f2da28b785b89e49b5d7c7cc7f8d133a04d9cfe76e1fb59 +size 904385440 diff --git a/tinyllama-1.1b-intermediate-step-1431k-3t.q8_0.gguf b/tinyllama-1.1b-intermediate-step-1431k-3t.q8_0.gguf new file mode 100644 index 0000000..a8ec5b4 --- /dev/null +++ b/tinyllama-1.1b-intermediate-step-1431k-3t.q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02cce05002a768b2d95904d289692c757e03dd35dd7f8d6a40566c30aedc33bb +size 1170781088