From 8be7019f39c25b7c5477cbe21472c050f1025abd Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sat, 2 May 2026 20:10:27 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF Source: Original Platform --- .gitattributes | 60 +++++++++++++++++++ README.md | 47 +++++++++++++++ configuration.json | 1 + featherless-quants.png | 3 + ...c-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf | 3 + sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf | 3 + ...c-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf | 3 + ...c-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf | 3 + ...c-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf | 3 + ...c-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf | 3 + ...c-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf | 3 + ...c-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf | 3 + ...c-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf | 3 + sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf | 3 + sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf | 3 + 15 files changed, 144 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 configuration.json create mode 100644 featherless-quants.png create mode 100644 sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf create mode 100644 sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf create mode 100644 sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf create mode 100644 sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf create mode 100644 sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf create mode 100644 sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf create mode 100644 sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf create mode 100644 sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf create mode 100644 sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf create mode 100644 sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf create mode 100644 sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..7815bbc --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bin.* filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zstandard filter=lfs diff=lfs merge=lfs -text +*.tfevents* filter=lfs diff=lfs merge=lfs -text +*.db* filter=lfs diff=lfs merge=lfs -text +*.ark* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text + +*.ggml filter=lfs diff=lfs merge=lfs -text +*.llamafile* filter=lfs diff=lfs merge=lfs -text +*.pt2 filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text + +sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +featherless-quants.png filter=lfs diff=lfs merge=lfs -text +sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text \ No newline at end of file diff --git a/README.md b/README.md new file mode 100644 index 0000000..e4d06a2 --- /dev/null +++ b/README.md @@ -0,0 +1,47 @@ +--- +base_model: sh2orc/Llama-3.1-Korean-8B-Instruct +pipeline_tag: text-generation +quantized_by: featherless-ai-quants +--- + +# sh2orc/Llama-3.1-Korean-8B-Instruct GGUF Quantizations 🚀 + +![Featherless AI Quants](./featherless-quants.png) + +*Optimized GGUF quantization files for enhanced model performance* + +> Powered by [Featherless AI](https://featherless.ai) - run any model you'd like for a simple small fee. +--- + +## Available Quantizations 📊 + +| Quantization Type | File | Size | +|-------------------|------|------| +| IQ4_XS | [sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf) | 4276.63 MB | +| Q2_K | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf) | 3031.86 MB | +| Q3_K_L | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf) | 4121.74 MB | +| Q3_K_M | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf) | 3832.74 MB | +| Q3_K_S | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf) | 3494.74 MB | +| Q4_K_M | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf) | 4692.78 MB | +| Q4_K_S | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf) | 4475.28 MB | +| Q5_K_M | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf) | 5467.41 MB | +| Q5_K_S | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf) | 5339.91 MB | +| Q6_K | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf) | 6290.45 MB | +| Q8_0 | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf) | 8145.12 MB | + + +--- + +## ⚡ Powered by [Featherless AI](https://featherless.ai) + +### Key Features + +- 🔥 **Instant Hosting** - Deploy any Llama model on HuggingFace instantly +- 🛠️ **Zero Infrastructure** - No server setup or maintenance required +- 📚 **Vast Compatibility** - Support for 2400+ models and counting +- 💎 **Affordable Pricing** - Starting at just $10/month + +--- + +**Links:** +[Get Started](https://featherless.ai) | [Documentation](https://featherless.ai/docs) | [Models](https://featherless.ai/models) \ No newline at end of file diff --git a/configuration.json b/configuration.json new file mode 100644 index 0000000..bbeeda1 --- /dev/null +++ b/configuration.json @@ -0,0 +1 @@ +{"framework": "pytorch", "task": "text-generation", "allow_remote": true} \ No newline at end of file diff --git a/featherless-quants.png b/featherless-quants.png new file mode 100644 index 0000000..3e0eab9 --- /dev/null +++ b/featherless-quants.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e1b4d66c8306c7b0614089381fdf86ea4efb02dffb78d22767a084cb8b88d6b +size 1614532 diff --git a/sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf b/sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf new file mode 100644 index 0000000..98a6ab0 --- /dev/null +++ b/sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8cf5e4612789fec8cb8ef0ace57bde66b9431cdc4c3036c7174fe36c3ad43c2 +size 4484367616 diff --git a/sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf new file mode 100644 index 0000000..fd6cb2f --- /dev/null +++ b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac304f8d6ff1738b14aa0cbb790ed8afc3a034ca4b26a94e8039a2a153efb107 +size 3179136256 diff --git a/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf new file mode 100644 index 0000000..b2c5e20 --- /dev/null +++ b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:221eb454dd7e7bc3d31c2a1ad01583de92b0fb9a3f60a18df61a20cb09b6e419 +size 4321961216 diff --git a/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf new file mode 100644 index 0000000..b626c83 --- /dev/null +++ b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:adf6d1a3f489a05434bf702d0e624356b4e35f8956e351dc5c58ddb1c8df0da1 +size 4018922752 diff --git a/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf new file mode 100644 index 0000000..042acdf --- /dev/null +++ b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e85c6b6956d88a03408c2c5d3145c557b2029dc3f29a6f86236d76a3e3fcc32 +size 3664504064 diff --git a/sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf new file mode 100644 index 0000000..459c3e5 --- /dev/null +++ b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66783621229735cdfabd45fc896fcf2909fc04d385eb84112fa6c5c406f4cc28 +size 4920739072 diff --git a/sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf new file mode 100644 index 0000000..a3dcd41 --- /dev/null +++ b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f84bd2990523deb274d8accdf9eb6e4ee09d596ca7208a97e88839cbef35d8d9 +size 4692673792 diff --git a/sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf new file mode 100644 index 0000000..7c9056a --- /dev/null +++ b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19a7c911444e01c02a3f37d89b92ccd8cb54949512c6d309fe80882d94b3e920 +size 5732992256 diff --git a/sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf new file mode 100644 index 0000000..fd4f949 --- /dev/null +++ b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66a7537e4005cd9a5398b5e43602468baa3391307606adff6e331ca74c081049 +size 5599298816 diff --git a/sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf new file mode 100644 index 0000000..fd4b4d8 --- /dev/null +++ b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b30f9cf0eefac895853f82584a49df05b201e60ed403bf41daf2ac282a7e86c +size 6596011264 diff --git a/sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf new file mode 100644 index 0000000..7cd8ccd --- /dev/null +++ b/sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aaa9d0a4be911840a83e05f6ef43b24fa3e81603b4a72def2ef73e4955a7c5be +size 8540775680