From 23f9fa2ff3fbe0464865c7c824960dd5d22f1be4 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 28 Apr 2026 22:57:42 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: afrideva/smol_llama-220M-bees-internal-GGUF Source: Original Platform --- .gitattributes | 42 +++++++ README.md | 141 ++++++++++++++++++++++ smol_llama-220m-bees-internal.fp16.gguf | 3 + smol_llama-220m-bees-internal.q2_k.gguf | 3 + smol_llama-220m-bees-internal.q3_k_m.gguf | 3 + smol_llama-220m-bees-internal.q4_k_m.gguf | 3 + smol_llama-220m-bees-internal.q5_k_m.gguf | 3 + smol_llama-220m-bees-internal.q6_k.gguf | 3 + smol_llama-220m-bees-internal.q8_0.gguf | 3 + 9 files changed, 204 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 smol_llama-220m-bees-internal.fp16.gguf create mode 100644 smol_llama-220m-bees-internal.q2_k.gguf create mode 100644 smol_llama-220m-bees-internal.q3_k_m.gguf create mode 100644 smol_llama-220m-bees-internal.q4_k_m.gguf create mode 100644 smol_llama-220m-bees-internal.q5_k_m.gguf create mode 100644 smol_llama-220m-bees-internal.q6_k.gguf create mode 100644 smol_llama-220m-bees-internal.q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..af8df74 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,42 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-bees-internal.fp16.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-bees-internal.q2_k.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-bees-internal.q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-bees-internal.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-bees-internal.q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-bees-internal.q6_k.gguf filter=lfs diff=lfs merge=lfs -text +smol_llama-220m-bees-internal.q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..36da462 --- /dev/null +++ b/README.md @@ -0,0 +1,141 @@ +--- +base_model: BEE-spoke-data/smol_llama-220M-bees-internal +datasets: +- BEE-spoke-data/bees-internal +inference: false +language: +- en +license: apache-2.0 +metrics: +- accuracy +model_creator: BEE-spoke-data +model_name: smol_llama-220M-bees-internal +pipeline_tag: text-generation +quantized_by: afrideva +tags: +- generated_from_trainer +- gguf +- ggml +- quantized +- q2_k +- q3_k_m +- q4_k_m +- q5_k_m +- q6_k +- q8_0 +widget: +- example_title: Queen Excluder + text: In beekeeping, the term "queen excluder" refers to +- example_title: Increasing Honey Production + text: One way to encourage a honey bee colony to produce more honey is by +- example_title: Lifecycle of a Worker Bee + text: The lifecycle of a worker bee consists of several stages, starting with +- example_title: Varroa Destructor + text: Varroa destructor is a type of mite that +- example_title: Beekeeping PPE + text: In the world of beekeeping, the acronym PPE stands for +- example_title: Robbing in Beekeeping + text: The term "robbing" in beekeeping refers to the act of +- example_title: Role of Drone Bees + text: 'Question: What''s the primary function of drone bees in a hive? + + Answer:' +- example_title: Honey Harvesting Device + text: To harvest honey from a hive, beekeepers often use a device known as a +- example_title: Beekeeping Math Problem + text: 'Problem: You have a hive that produces 60 pounds of honey per year. You decide + to split the hive into two. Assuming each hive now produces at a 70% rate compared + to before, how much honey will you get from both hives next year? + + To calculate' +- example_title: Swarming + text: In beekeeping, "swarming" is the process where +--- +# BEE-spoke-data/smol_llama-220M-bees-internal-GGUF + +Quantized GGUF model files for [smol_llama-220M-bees-internal](https://huggingface.co/BEE-spoke-data/smol_llama-220M-bees-internal) from [BEE-spoke-data](https://huggingface.co/BEE-spoke-data) + + +| Name | Quant method | Size | +| ---- | ---- | ---- | +| [smol_llama-220m-bees-internal.fp16.gguf](https://huggingface.co/afrideva/smol_llama-220M-bees-internal-GGUF/resolve/main/smol_llama-220m-bees-internal.fp16.gguf) | fp16 | 436.50 MB | +| [smol_llama-220m-bees-internal.q2_k.gguf](https://huggingface.co/afrideva/smol_llama-220M-bees-internal-GGUF/resolve/main/smol_llama-220m-bees-internal.q2_k.gguf) | q2_k | 94.43 MB | +| [smol_llama-220m-bees-internal.q3_k_m.gguf](https://huggingface.co/afrideva/smol_llama-220M-bees-internal-GGUF/resolve/main/smol_llama-220m-bees-internal.q3_k_m.gguf) | q3_k_m | 114.65 MB | +| [smol_llama-220m-bees-internal.q4_k_m.gguf](https://huggingface.co/afrideva/smol_llama-220M-bees-internal-GGUF/resolve/main/smol_llama-220m-bees-internal.q4_k_m.gguf) | q4_k_m | 137.58 MB | +| [smol_llama-220m-bees-internal.q5_k_m.gguf](https://huggingface.co/afrideva/smol_llama-220M-bees-internal-GGUF/resolve/main/smol_llama-220m-bees-internal.q5_k_m.gguf) | q5_k_m | 157.91 MB | +| [smol_llama-220m-bees-internal.q6_k.gguf](https://huggingface.co/afrideva/smol_llama-220M-bees-internal-GGUF/resolve/main/smol_llama-220m-bees-internal.q6_k.gguf) | q6_k | 179.52 MB | +| [smol_llama-220m-bees-internal.q8_0.gguf](https://huggingface.co/afrideva/smol_llama-220M-bees-internal-GGUF/resolve/main/smol_llama-220m-bees-internal.q8_0.gguf) | q8_0 | 232.28 MB | + + + +## Original Model Card: + + +# smol_llama-220M-bees-internal + +This model is a fine-tuned version of [BEE-spoke-data/smol_llama-220M-GQA](https://huggingface.co/BEE-spoke-data/smol_llama-220M-GQA) on the None dataset. +It achieves the following results on the evaluation set: +- Loss: 2.6892 +- Accuracy: 0.4610 + +## Model description + +More information needed + +## Intended uses & limitations + +More information needed + +## Training and evaluation data + +More information needed + +## Training procedure + +### Training hyperparameters + +The following hyperparameters were used during training: +- learning_rate: 0.0001 +- train_batch_size: 4 +- eval_batch_size: 2 +- seed: 27634 +- gradient_accumulation_steps: 8 +- total_train_batch_size: 32 +- optimizer: Adam with betas=(0.9,0.95) and epsilon=1e-08 +- lr_scheduler_type: cosine +- lr_scheduler_warmup_ratio: 0.05 +- num_epochs: 2.0 + +### Training results + +| Training Loss | Epoch | Step | Validation Loss | Accuracy | +|:-------------:|:-----:|:----:|:---------------:|:--------:| +| 3.0959 | 0.1 | 50 | 2.9671 | 0.4245 | +| 2.9975 | 0.19 | 100 | 2.8691 | 0.4371 | +| 2.8938 | 0.29 | 150 | 2.8271 | 0.4419 | +| 2.9027 | 0.39 | 200 | 2.7973 | 0.4457 | +| 2.8983 | 0.49 | 250 | 2.7719 | 0.4489 | +| 2.8789 | 0.58 | 300 | 2.7519 | 0.4515 | +| 2.8672 | 0.68 | 350 | 2.7366 | 0.4535 | +| 2.8369 | 0.78 | 400 | 2.7230 | 0.4558 | +| 2.8271 | 0.88 | 450 | 2.7118 | 0.4569 | +| 2.7775 | 0.97 | 500 | 2.7034 | 0.4587 | +| 2.671 | 1.07 | 550 | 2.6996 | 0.4592 | +| 2.695 | 1.17 | 600 | 2.6965 | 0.4598 | +| 2.6962 | 1.27 | 650 | 2.6934 | 0.4601 | +| 2.6034 | 1.36 | 700 | 2.6916 | 0.4605 | +| 2.716 | 1.46 | 750 | 2.6901 | 0.4609 | +| 2.6968 | 1.56 | 800 | 2.6896 | 0.4608 | +| 2.6626 | 1.66 | 850 | 2.6893 | 0.4609 | +| 2.6881 | 1.75 | 900 | 2.6891 | 0.4610 | +| 2.7339 | 1.85 | 950 | 2.6891 | 0.4610 | +| 2.6729 | 1.95 | 1000 | 2.6892 | 0.4610 | + + +### Framework versions + +- Transformers 4.36.2 +- Pytorch 2.1.0 +- Datasets 2.16.1 +- Tokenizers 0.15.0 \ No newline at end of file diff --git a/smol_llama-220m-bees-internal.fp16.gguf b/smol_llama-220m-bees-internal.fp16.gguf new file mode 100644 index 0000000..327bbf9 --- /dev/null +++ b/smol_llama-220m-bees-internal.fp16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9fd00768d60601242469d1b2e09b35dae94e68b41ec0a2180b00044d1aa7f7d +size 436502176 diff --git a/smol_llama-220m-bees-internal.q2_k.gguf b/smol_llama-220m-bees-internal.q2_k.gguf new file mode 100644 index 0000000..5e0abd1 --- /dev/null +++ b/smol_llama-220m-bees-internal.q2_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f891efb9c6838847eca6907405f84b7b262e8cf9143ce49d62055b5bfbaf3381 +size 94429888 diff --git a/smol_llama-220m-bees-internal.q3_k_m.gguf b/smol_llama-220m-bees-internal.q3_k_m.gguf new file mode 100644 index 0000000..e029e0d --- /dev/null +++ b/smol_llama-220m-bees-internal.q3_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa5b309c2ef7503f70181aaa0d39054b4971ccb107ef8faecdf874c759f124b5 +size 114650816 diff --git a/smol_llama-220m-bees-internal.q4_k_m.gguf b/smol_llama-220m-bees-internal.q4_k_m.gguf new file mode 100644 index 0000000..9af6274 --- /dev/null +++ b/smol_llama-220m-bees-internal.q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a95c237c3cfc81abaefe6243077f9cb860a9618773c71a0b733b09afb7367de +size 137581248 diff --git a/smol_llama-220m-bees-internal.q5_k_m.gguf b/smol_llama-220m-bees-internal.q5_k_m.gguf new file mode 100644 index 0000000..5d4538f --- /dev/null +++ b/smol_llama-220m-bees-internal.q5_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d56e66d0029a9353909c4f8c9551951079d47102e4eeda173e5ab624076bdb99 +size 157913792 diff --git a/smol_llama-220m-bees-internal.q6_k.gguf b/smol_llama-220m-bees-internal.q6_k.gguf new file mode 100644 index 0000000..fc69088 --- /dev/null +++ b/smol_llama-220m-bees-internal.q6_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05b2fd6885e2137a0ba159c46110e473382149d2627d333498c49fcd7b79223f +size 179517120 diff --git a/smol_llama-220m-bees-internal.q8_0.gguf b/smol_llama-220m-bees-internal.q8_0.gguf new file mode 100644 index 0000000..9cb856b --- /dev/null +++ b/smol_llama-220m-bees-internal.q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:002f195802bbc9fcbf425dc91d94c4dbf3c54ff04b30c83c83302b3f3925588f +size 232275648