From b458b301aaf884177cccdb6b393c8a41e083476f Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Fri, 1 May 2026 16:49:10 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: akumaburn/Alpaca-Llama-3-8B-GGUF Source: Original Platform --- .gitattributes | 42 ++++++++++++ Alpaca-Llama-3-8B-GGUF-unsloth.F16.gguf | 3 + Alpaca-Llama-3-8B-GGUF-unsloth.Q2_K.gguf | 3 + Alpaca-Llama-3-8B-GGUF-unsloth.Q3_K_M.gguf | 3 + Alpaca-Llama-3-8B-GGUF-unsloth.Q4_K_M.gguf | 3 + Alpaca-Llama-3-8B-GGUF-unsloth.Q5_K_M.gguf | 3 + Alpaca-Llama-3-8B-GGUF-unsloth.Q6_K.gguf | 3 + Alpaca-Llama-3-8B-GGUF-unsloth.Q8_0.gguf | 3 + README.md | 75 ++++++++++++++++++++++ config.json | 3 + 10 files changed, 141 insertions(+) create mode 100644 .gitattributes create mode 100644 Alpaca-Llama-3-8B-GGUF-unsloth.F16.gguf create mode 100644 Alpaca-Llama-3-8B-GGUF-unsloth.Q2_K.gguf create mode 100644 Alpaca-Llama-3-8B-GGUF-unsloth.Q3_K_M.gguf create mode 100644 Alpaca-Llama-3-8B-GGUF-unsloth.Q4_K_M.gguf create mode 100644 Alpaca-Llama-3-8B-GGUF-unsloth.Q5_K_M.gguf create mode 100644 Alpaca-Llama-3-8B-GGUF-unsloth.Q6_K.gguf create mode 100644 Alpaca-Llama-3-8B-GGUF-unsloth.Q8_0.gguf create mode 100644 README.md create mode 100644 config.json diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..13482eb --- /dev/null +++ b/.gitattributes @@ -0,0 +1,42 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Alpaca-Llama-3-8B-GGUF-unsloth.F16.gguf filter=lfs diff=lfs merge=lfs -text +Alpaca-Llama-3-8B-GGUF-unsloth.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Alpaca-Llama-3-8B-GGUF-unsloth.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Alpaca-Llama-3-8B-GGUF-unsloth.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Alpaca-Llama-3-8B-GGUF-unsloth.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Alpaca-Llama-3-8B-GGUF-unsloth.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Alpaca-Llama-3-8B-GGUF-unsloth.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Alpaca-Llama-3-8B-GGUF-unsloth.F16.gguf b/Alpaca-Llama-3-8B-GGUF-unsloth.F16.gguf new file mode 100644 index 0000000..b3e2610 --- /dev/null +++ b/Alpaca-Llama-3-8B-GGUF-unsloth.F16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c83bba6559b972f5567ec9336cc66cf111aad54cc90407810b3dcf7a494d8df5 +size 16069403488 diff --git a/Alpaca-Llama-3-8B-GGUF-unsloth.Q2_K.gguf b/Alpaca-Llama-3-8B-GGUF-unsloth.Q2_K.gguf new file mode 100644 index 0000000..5a8caa2 --- /dev/null +++ b/Alpaca-Llama-3-8B-GGUF-unsloth.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6db8e60767d6e50c09c612a0be092f7ccbe91c4d9ff43ab3f388ccaa258ee8e5 +size 3179643776 diff --git a/Alpaca-Llama-3-8B-GGUF-unsloth.Q3_K_M.gguf b/Alpaca-Llama-3-8B-GGUF-unsloth.Q3_K_M.gguf new file mode 100644 index 0000000..26bef57 --- /dev/null +++ b/Alpaca-Llama-3-8B-GGUF-unsloth.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:615de448a2be4688d51361b7c80a35ce515c60c50f243babb7355a9b5db88037 +size 4019430272 diff --git a/Alpaca-Llama-3-8B-GGUF-unsloth.Q4_K_M.gguf b/Alpaca-Llama-3-8B-GGUF-unsloth.Q4_K_M.gguf new file mode 100644 index 0000000..f0944e7 --- /dev/null +++ b/Alpaca-Llama-3-8B-GGUF-unsloth.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:024e514da40f83eeaffbacfab4d13316a4a8dae998ae8b94f3db54ab22338350 +size 4921246592 diff --git a/Alpaca-Llama-3-8B-GGUF-unsloth.Q5_K_M.gguf b/Alpaca-Llama-3-8B-GGUF-unsloth.Q5_K_M.gguf new file mode 100644 index 0000000..3551db8 --- /dev/null +++ b/Alpaca-Llama-3-8B-GGUF-unsloth.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce070e0c71438bed39b76051f8a3ff2932b118f64569cffe2ddf1e54eb292b24 +size 5733499776 diff --git a/Alpaca-Llama-3-8B-GGUF-unsloth.Q6_K.gguf b/Alpaca-Llama-3-8B-GGUF-unsloth.Q6_K.gguf new file mode 100644 index 0000000..ebd8fc3 --- /dev/null +++ b/Alpaca-Llama-3-8B-GGUF-unsloth.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77b33ed3396c6ea37ede7774b0b827bcb9bbffd198dba53286d4b9046e44b601 +size 6596518784 diff --git a/Alpaca-Llama-3-8B-GGUF-unsloth.Q8_0.gguf b/Alpaca-Llama-3-8B-GGUF-unsloth.Q8_0.gguf new file mode 100644 index 0000000..ffc5f67 --- /dev/null +++ b/Alpaca-Llama-3-8B-GGUF-unsloth.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2eb2f24a5d637bf4c8aefcb3d18c876362f90267c7c34db29c38fdd2488a2b4d +size 8541283168 diff --git a/README.md b/README.md new file mode 100644 index 0000000..3704afa --- /dev/null +++ b/README.md @@ -0,0 +1,75 @@ +--- +language: +- en +license: apache-2.0 +tags: +- text-generation-inference +- transformers +- unsloth +- llama +- trl +- sft +base_model: unsloth/llama-3-8b-bnb-4bit +datasets: +- yahma/alpaca-cleaned +--- + +# Alpaca-Llama-3-8B + + +- **Fine Tuned using dataset:** https://huggingface.co/datasets/yahma/alpaca-cleaned +- **Epoch Count:** 1 +- **Step Count:** 6,470/6,470 +- **Batch Size:** 2 +- **Gradient Accumulation Steps:** 4 +- **Context Size:** 8192 +- **Num examples:** 51,760 +- **Trainable Parameters:** 41,943,040 +- **Learning Rate:** 0.00001 +- **Training Loss:** 0.960000 +- **Fined Tuned using:** Google Colab Pro (Nvidia T4 runtime) + +- **Developed by:** akumaburn +- **License:** apache-2.0 +- **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit +- **Prompt Format:** Alpaca (https://libertai.io/apis/text-generation/prompting.html) +- **Chai ELO:** 1146.84 (https://console.chaiverse.com/models/akumaburn-alpaca-llama-3-8b_v1) + +Full model can be found in https://huggingface.co/akumaburn/Alpaca-Llama-3-8B + +mistral-7b-openorca.Q8_0.gguf: +- **MMLU-Test:** Final result: **41.5836 +/- 0.4174** +- **Arc-Easy:** Final result: 72.6316 +/- 1.8691 +- **Truthful QA:** Final result: **32.0685 +/- 1.6339** +- **Arc-Challenge:** Final result: 48.8294 +/- 2.8956 + +llama-3-8b-bnb-4bit.Q8_0.gguf: +- **MMLU-Test:** Final result: 40.4074 +/- 0.4156 +- **Arc-Easy:** Final result: 73.8596 +/- 1.8421 +- **Truthful QA:** Final result: 26.6830 +/- 1.5484 +- **Arc-Challenge:** Final result: 46.8227 +/- 2.8906 + +Open_Orca_Llama-3-8B-unsloth.Q8_0.gguf: +- **MMLU-Test:** Final result: 39.3818 +/- 0.4138 +- **Arc-Easy:** Final result: 67.3684 +/- 1.9656 +- **Truthful QA:** Final result: 29.0086 +/- 1.5886 +- **Arc-Challenge:** Final result: 42.1405 +/- 2.8604 + +**Alpaca-Llama-3-8B-GGUF-unsloth.Q8_0.gguf**: +- **MMLU-Test:** Final result: 40.6441 +/- 0.4160 +- **Arc-Easy:** Final result: **77.5439 +/- 1.7494** +- **Truthful QA:** Final result: 29.7430 +/- 1.6003 +- **Arc-Challenge:** Final result: **50.5017 +/- 2.8963** + +Meta-Llama-3-8B.Q8_0.gguf: +- **MMLU-Test:** Final result: 40.8664 +/- 0.4163 +- **Arc-Easy:** Final result: 74.3860 +/- 1.8299 +- **Truthful QA:** Final result: 28.6414 +/- 1.5826 +- **Arc-Challenge:** Final result: 47.1572 +/- 2.8917 + +Llama.cpp Options For Testing: +--samplers "tfs;typical;temp" --draft 32 --ctx-size 8192 --temp 0.82 --tfs 0.8 --typical 1.1 --repeat-last-n 512 --batch-size 8192 --repeat-penalty 1.0 --n-gpu-layers 100 --threads 12 + +This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. + +[](https://github.com/unslothai/unsloth) diff --git a/config.json b/config.json new file mode 100644 index 0000000..a4ba21b --- /dev/null +++ b/config.json @@ -0,0 +1,3 @@ +{ + "model_type": "llama" +} \ No newline at end of file