From de7b87c55f30df792d73185efe72e536819700ec Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 28 Apr 2026 22:57:44 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: afrideva/TinyLlama-3T-1.1bee-GGUF Source: Original Platform --- .gitattributes | 42 ++++++++++ README.md | 132 ++++++++++++++++++++++++++++++++ tinyllama-3t-1.1bee.fp16.gguf | 3 + tinyllama-3t-1.1bee.q2_k.gguf | 3 + tinyllama-3t-1.1bee.q3_k_m.gguf | 3 + tinyllama-3t-1.1bee.q4_k_m.gguf | 3 + tinyllama-3t-1.1bee.q5_k_m.gguf | 3 + tinyllama-3t-1.1bee.q6_k.gguf | 3 + tinyllama-3t-1.1bee.q8_0.gguf | 3 + 9 files changed, 195 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 tinyllama-3t-1.1bee.fp16.gguf create mode 100644 tinyllama-3t-1.1bee.q2_k.gguf create mode 100644 tinyllama-3t-1.1bee.q3_k_m.gguf create mode 100644 tinyllama-3t-1.1bee.q4_k_m.gguf create mode 100644 tinyllama-3t-1.1bee.q5_k_m.gguf create mode 100644 tinyllama-3t-1.1bee.q6_k.gguf create mode 100644 tinyllama-3t-1.1bee.q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..559073f --- /dev/null +++ b/.gitattributes @@ -0,0 +1,42 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +tinyllama-3t-1.1bee.fp16.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-3t-1.1bee.q2_k.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-3t-1.1bee.q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-3t-1.1bee.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-3t-1.1bee.q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-3t-1.1bee.q6_k.gguf filter=lfs diff=lfs merge=lfs -text +tinyllama-3t-1.1bee.q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..698e354 --- /dev/null +++ b/README.md @@ -0,0 +1,132 @@ +--- +base_model: BEE-spoke-data/TinyLlama-3T-1.1bee +datasets: +- BEE-spoke-data/bees-internal +inference: false +language: +- en +license: apache-2.0 +metrics: +- accuracy +model_creator: BEE-spoke-data +model_name: TinyLlama-3T-1.1bee +pipeline_tag: text-generation +quantized_by: afrideva +tags: +- bees +- bzz +- honey +- oprah winfrey +- gguf +- ggml +- quantized +- q2_k +- q3_k_m +- q4_k_m +- q5_k_m +- q6_k +- q8_0 +widget: +- example_title: Queen Excluder + text: In beekeeping, the term "queen excluder" refers to +- example_title: Increasing Honey Production + text: One way to encourage a honey bee colony to produce more honey is by +- example_title: Lifecycle of a Worker Bee + text: The lifecycle of a worker bee consists of several stages, starting with +- example_title: Varroa Destructor + text: Varroa destructor is a type of mite that +- example_title: Beekeeping PPE + text: In the world of beekeeping, the acronym PPE stands for +- example_title: Robbing in Beekeeping + text: The term "robbing" in beekeeping refers to the act of +- example_title: Role of Drone Bees + text: 'Question: What''s the primary function of drone bees in a hive? + + Answer:' +- example_title: Honey Harvesting Device + text: To harvest honey from a hive, beekeepers often use a device known as a +- example_title: Beekeeping Math Problem + text: 'Problem: You have a hive that produces 60 pounds of honey per year. You decide + to split the hive into two. Assuming each hive now produces at a 70% rate compared + to before, how much honey will you get from both hives next year? + + To calculate' +- example_title: Swarming + text: In beekeeping, "swarming" is the process where +--- +# BEE-spoke-data/TinyLlama-3T-1.1bee-GGUF + +Quantized GGUF model files for [TinyLlama-3T-1.1bee](https://huggingface.co/BEE-spoke-data/TinyLlama-3T-1.1bee) from [BEE-spoke-data](https://huggingface.co/BEE-spoke-data) + + +| Name | Quant method | Size | +| ---- | ---- | ---- | +| [tinyllama-3t-1.1bee.fp16.gguf](https://huggingface.co/afrideva/TinyLlama-3T-1.1bee-GGUF/resolve/main/tinyllama-3t-1.1bee.fp16.gguf) | fp16 | 2.20 GB | +| [tinyllama-3t-1.1bee.q2_k.gguf](https://huggingface.co/afrideva/TinyLlama-3T-1.1bee-GGUF/resolve/main/tinyllama-3t-1.1bee.q2_k.gguf) | q2_k | 432.13 MB | +| [tinyllama-3t-1.1bee.q3_k_m.gguf](https://huggingface.co/afrideva/TinyLlama-3T-1.1bee-GGUF/resolve/main/tinyllama-3t-1.1bee.q3_k_m.gguf) | q3_k_m | 548.40 MB | +| [tinyllama-3t-1.1bee.q4_k_m.gguf](https://huggingface.co/afrideva/TinyLlama-3T-1.1bee-GGUF/resolve/main/tinyllama-3t-1.1bee.q4_k_m.gguf) | q4_k_m | 667.81 MB | +| [tinyllama-3t-1.1bee.q5_k_m.gguf](https://huggingface.co/afrideva/TinyLlama-3T-1.1bee-GGUF/resolve/main/tinyllama-3t-1.1bee.q5_k_m.gguf) | q5_k_m | 782.04 MB | +| [tinyllama-3t-1.1bee.q6_k.gguf](https://huggingface.co/afrideva/TinyLlama-3T-1.1bee-GGUF/resolve/main/tinyllama-3t-1.1bee.q6_k.gguf) | q6_k | 903.41 MB | +| [tinyllama-3t-1.1bee.q8_0.gguf](https://huggingface.co/afrideva/TinyLlama-3T-1.1bee-GGUF/resolve/main/tinyllama-3t-1.1bee.q8_0.gguf) | q8_0 | 1.17 GB | + + + +## Original Model Card: + + +# TinyLlama-3T-1.1bee + + +![image/png](https://cdn-uploads.huggingface.co/production/uploads/60bccec062080d33f875cd0c/I6AfPId0Xo_vVobtkAP12.png) + +A grand successor to [the original](https://huggingface.co/BEE-spoke-data/TinyLlama-1.1bee). This one has the following improvements: + +- start from [finished 3T TinyLlama](https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T) +- vastly improved and expanded SoTA beekeeping dataset + + +## Model description + +This model is a fine-tuned version of TinyLlama-1.1b-3T on the BEE-spoke-data/bees-internal dataset. + +It achieves the following results on the evaluation set: +- Loss: 2.1640 +- Accuracy: 0.5406 + +### Training hyperparameters + +The following hyperparameters were used during training: +- learning_rate: 0.0001 +- train_batch_size: 4 +- eval_batch_size: 2 +- seed: 13707 +- gradient_accumulation_steps: 16 +- total_train_batch_size: 64 +- optimizer: Adam with betas=(0.9,0.95) and epsilon=1e-08 +- lr_scheduler_type: cosine +- lr_scheduler_warmup_ratio: 0.05 +- num_epochs: 2.0 + +### Training results + +| Training Loss | Epoch | Step | Validation Loss | Accuracy | +|:-------------:|:-----:|:----:|:---------------:|:--------:| +| 2.4432 | 0.19 | 50 | 2.3850 | 0.5033 | +| 2.3655 | 0.39 | 100 | 2.3124 | 0.5129 | +| 2.374 | 0.58 | 150 | 2.2588 | 0.5215 | +| 2.3558 | 0.78 | 200 | 2.2132 | 0.5291 | +| 2.2677 | 0.97 | 250 | 2.1828 | 0.5348 | +| 2.0701 | 1.17 | 300 | 2.1788 | 0.5373 | +| 2.0766 | 1.36 | 350 | 2.1673 | 0.5398 | +| 2.0669 | 1.56 | 400 | 2.1651 | 0.5402 | +| 2.0314 | 1.75 | 450 | 2.1641 | 0.5406 | +| 2.0281 | 1.95 | 500 | 2.1639 | 0.5407 | + + +### Framework versions + +- Transformers 4.36.2 +- Pytorch 2.1.0 +- Datasets 2.16.1 +- Tokenizers 0.15.0 \ No newline at end of file diff --git a/tinyllama-3t-1.1bee.fp16.gguf b/tinyllama-3t-1.1bee.fp16.gguf new file mode 100644 index 0000000..7dcfbf1 --- /dev/null +++ b/tinyllama-3t-1.1bee.fp16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92dc410393b4eaada50246100f573d5ab4802734133e50cd4683d2ddae38adbb +size 2201017248 diff --git a/tinyllama-3t-1.1bee.q2_k.gguf b/tinyllama-3t-1.1bee.q2_k.gguf new file mode 100644 index 0000000..d9c23f1 --- /dev/null +++ b/tinyllama-3t-1.1bee.q2_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4668884aaf634acb6b09710eb26a1ad283a4648a70056173748edd67d087551b +size 432131008 diff --git a/tinyllama-3t-1.1bee.q3_k_m.gguf b/tinyllama-3t-1.1bee.q3_k_m.gguf new file mode 100644 index 0000000..7abe986 --- /dev/null +++ b/tinyllama-3t-1.1bee.q3_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b2d35c1df09741c18d69b0eda699e50f5824dfc9258c261b0abae468cb64ed9 +size 548404160 diff --git a/tinyllama-3t-1.1bee.q4_k_m.gguf b/tinyllama-3t-1.1bee.q4_k_m.gguf new file mode 100644 index 0000000..5c55328 --- /dev/null +++ b/tinyllama-3t-1.1bee.q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3f28a88c103a890a68bf8d437003a80b80a51082120718ed646786b7a1a7b5e +size 667814848 diff --git a/tinyllama-3t-1.1bee.q5_k_m.gguf b/tinyllama-3t-1.1bee.q5_k_m.gguf new file mode 100644 index 0000000..7bae278 --- /dev/null +++ b/tinyllama-3t-1.1bee.q5_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf1902cce7090d86fdc71b864531ed35d792b7f3d68a25283179852908347fa3 +size 782044096 diff --git a/tinyllama-3t-1.1bee.q6_k.gguf b/tinyllama-3t-1.1bee.q6_k.gguf new file mode 100644 index 0000000..a04224f --- /dev/null +++ b/tinyllama-3t-1.1bee.q6_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ec8247028d27f018deffa958f9bcfffa350ca620a9c45cddf669e237a1faf6a +size 903412672 diff --git a/tinyllama-3t-1.1bee.q8_0.gguf b/tinyllama-3t-1.1bee.q8_0.gguf new file mode 100644 index 0000000..b79eab0 --- /dev/null +++ b/tinyllama-3t-1.1bee.q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e7ef0138d8d2020013ba0706c2bccd6d75dd7d2d6c15e9f828d59c7c5b6c703 +size 1169808320