初始化项目,由ModelHub XC社区提供模型
Model: featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF Source: Original Platform
This commit is contained in:
60
.gitattributes
vendored
Normal file
60
.gitattributes
vendored
Normal file
@@ -0,0 +1,60 @@
|
||||
*.7z filter=lfs diff=lfs merge=lfs -text
|
||||
*.arrow filter=lfs diff=lfs merge=lfs -text
|
||||
*.bin filter=lfs diff=lfs merge=lfs -text
|
||||
*.bin.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
||||
*.ftz filter=lfs diff=lfs merge=lfs -text
|
||||
*.gz filter=lfs diff=lfs merge=lfs -text
|
||||
*.h5 filter=lfs diff=lfs merge=lfs -text
|
||||
*.joblib filter=lfs diff=lfs merge=lfs -text
|
||||
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.model filter=lfs diff=lfs merge=lfs -text
|
||||
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
||||
*.onnx filter=lfs diff=lfs merge=lfs -text
|
||||
*.ot filter=lfs diff=lfs merge=lfs -text
|
||||
*.parquet filter=lfs diff=lfs merge=lfs -text
|
||||
*.pb filter=lfs diff=lfs merge=lfs -text
|
||||
*.pt filter=lfs diff=lfs merge=lfs -text
|
||||
*.pth filter=lfs diff=lfs merge=lfs -text
|
||||
*.rar filter=lfs diff=lfs merge=lfs -text
|
||||
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tflite filter=lfs diff=lfs merge=lfs -text
|
||||
*.tgz filter=lfs diff=lfs merge=lfs -text
|
||||
*.xz filter=lfs diff=lfs merge=lfs -text
|
||||
*.zip filter=lfs diff=lfs merge=lfs -text
|
||||
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
||||
*.tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||
*.db* filter=lfs diff=lfs merge=lfs -text
|
||||
*.ark* filter=lfs diff=lfs merge=lfs -text
|
||||
**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text
|
||||
**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text
|
||||
**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text
|
||||
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
||||
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
||||
|
||||
*.ggml filter=lfs diff=lfs merge=lfs -text
|
||||
*.llamafile* filter=lfs diff=lfs merge=lfs -text
|
||||
*.pt2 filter=lfs diff=lfs merge=lfs -text
|
||||
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
||||
*.npy filter=lfs diff=lfs merge=lfs -text
|
||||
*.npz filter=lfs diff=lfs merge=lfs -text
|
||||
*.pickle filter=lfs diff=lfs merge=lfs -text
|
||||
*.pkl filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar filter=lfs diff=lfs merge=lfs -text
|
||||
*.wasm filter=lfs diff=lfs merge=lfs -text
|
||||
*.zst filter=lfs diff=lfs merge=lfs -text
|
||||
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||
|
||||
sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
featherless-quants.png filter=lfs diff=lfs merge=lfs -text
|
||||
sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
47
README.md
Normal file
47
README.md
Normal file
@@ -0,0 +1,47 @@
|
||||
---
|
||||
base_model: sh2orc/Llama-3.1-Korean-8B-Instruct
|
||||
pipeline_tag: text-generation
|
||||
quantized_by: featherless-ai-quants
|
||||
---
|
||||
|
||||
# sh2orc/Llama-3.1-Korean-8B-Instruct GGUF Quantizations 🚀
|
||||
|
||||

|
||||
|
||||
*Optimized GGUF quantization files for enhanced model performance*
|
||||
|
||||
> Powered by [Featherless AI](https://featherless.ai) - run any model you'd like for a simple small fee.
|
||||
---
|
||||
|
||||
## Available Quantizations 📊
|
||||
|
||||
| Quantization Type | File | Size |
|
||||
|-------------------|------|------|
|
||||
| IQ4_XS | [sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf) | 4276.63 MB |
|
||||
| Q2_K | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf) | 3031.86 MB |
|
||||
| Q3_K_L | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf) | 4121.74 MB |
|
||||
| Q3_K_M | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf) | 3832.74 MB |
|
||||
| Q3_K_S | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf) | 3494.74 MB |
|
||||
| Q4_K_M | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf) | 4692.78 MB |
|
||||
| Q4_K_S | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf) | 4475.28 MB |
|
||||
| Q5_K_M | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf) | 5467.41 MB |
|
||||
| Q5_K_S | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf) | 5339.91 MB |
|
||||
| Q6_K | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf) | 6290.45 MB |
|
||||
| Q8_0 | [sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf](https://huggingface.co/featherless-ai-quants/sh2orc-Llama-3.1-Korean-8B-Instruct-GGUF/blob/main/sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf) | 8145.12 MB |
|
||||
|
||||
|
||||
---
|
||||
|
||||
## ⚡ Powered by [Featherless AI](https://featherless.ai)
|
||||
|
||||
### Key Features
|
||||
|
||||
- 🔥 **Instant Hosting** - Deploy any Llama model on HuggingFace instantly
|
||||
- 🛠️ **Zero Infrastructure** - No server setup or maintenance required
|
||||
- 📚 **Vast Compatibility** - Support for 2400+ models and counting
|
||||
- 💎 **Affordable Pricing** - Starting at just $10/month
|
||||
|
||||
---
|
||||
|
||||
**Links:**
|
||||
[Get Started](https://featherless.ai) | [Documentation](https://featherless.ai/docs) | [Models](https://featherless.ai/models)
|
||||
1
configuration.json
Normal file
1
configuration.json
Normal file
@@ -0,0 +1 @@
|
||||
{"framework": "pytorch", "task": "text-generation", "allow_remote": true}
|
||||
3
featherless-quants.png
Normal file
3
featherless-quants.png
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:2e1b4d66c8306c7b0614089381fdf86ea4efb02dffb78d22767a084cb8b88d6b
|
||||
size 1614532
|
||||
3
sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf
Normal file
3
sh2orc-Llama-3.1-Korean-8B-Instruct-IQ4_XS.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:d8cf5e4612789fec8cb8ef0ace57bde66b9431cdc4c3036c7174fe36c3ad43c2
|
||||
size 4484367616
|
||||
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf
Normal file
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q2_K.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:ac304f8d6ff1738b14aa0cbb790ed8afc3a034ca4b26a94e8039a2a153efb107
|
||||
size 3179136256
|
||||
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf
Normal file
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_L.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:221eb454dd7e7bc3d31c2a1ad01583de92b0fb9a3f60a18df61a20cb09b6e419
|
||||
size 4321961216
|
||||
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf
Normal file
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:adf6d1a3f489a05434bf702d0e624356b4e35f8956e351dc5c58ddb1c8df0da1
|
||||
size 4018922752
|
||||
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf
Normal file
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q3_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:8e85c6b6956d88a03408c2c5d3145c557b2029dc3f29a6f86236d76a3e3fcc32
|
||||
size 3664504064
|
||||
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf
Normal file
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:66783621229735cdfabd45fc896fcf2909fc04d385eb84112fa6c5c406f4cc28
|
||||
size 4920739072
|
||||
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf
Normal file
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q4_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:f84bd2990523deb274d8accdf9eb6e4ee09d596ca7208a97e88839cbef35d8d9
|
||||
size 4692673792
|
||||
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf
Normal file
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:19a7c911444e01c02a3f37d89b92ccd8cb54949512c6d309fe80882d94b3e920
|
||||
size 5732992256
|
||||
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf
Normal file
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q5_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:66a7537e4005cd9a5398b5e43602468baa3391307606adff6e331ca74c081049
|
||||
size 5599298816
|
||||
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf
Normal file
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q6_K.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:0b30f9cf0eefac895853f82584a49df05b201e60ed403bf41daf2ac282a7e86c
|
||||
size 6596011264
|
||||
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf
Normal file
3
sh2orc-Llama-3.1-Korean-8B-Instruct-Q8_0.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:aaa9d0a4be911840a83e05f6ef43b24fa3e81603b4a72def2ef73e4955a7c5be
|
||||
size 8540775680
|
||||
Reference in New Issue
Block a user