commit ebd454496aaa29da7cb40aa81ba36c2bfcb723d6 Author: ModelHub XC Date: Tue Apr 21 18:15:27 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: featherless-ai-quants/williamlcn-3400_mcq-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..31b725c --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +featherless-quants.png filter=lfs diff=lfs merge=lfs -text +williamlcn-3400_mcq-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +williamlcn-3400_mcq-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +williamlcn-3400_mcq-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +williamlcn-3400_mcq-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +williamlcn-3400_mcq-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +williamlcn-3400_mcq-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +williamlcn-3400_mcq-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +williamlcn-3400_mcq-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +williamlcn-3400_mcq-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +williamlcn-3400_mcq-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +williamlcn-3400_mcq-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..7b811b0 --- /dev/null +++ b/README.md @@ -0,0 +1,47 @@ +--- +base_model: williamlcn/3400_mcq +pipeline_tag: text-generation +quantized_by: featherless-ai-quants +--- + +# williamlcn/3400_mcq GGUF Quantizations 🚀 + +![Featherless AI Quants](./featherless-quants.png) + +*Optimized GGUF quantization files for enhanced model performance* + +> Powered by [Featherless AI](https://featherless.ai) - run any model you'd like for a simple small fee. +--- + +## Available Quantizations 📊 + +| Quantization Type | File | Size | +|-------------------|------|------| +| IQ4_XS | [williamlcn-3400_mcq-IQ4_XS.gguf](https://huggingface.co/featherless-ai-quants/williamlcn-3400_mcq-GGUF/blob/main/williamlcn-3400_mcq-IQ4_XS.gguf) | 1503.19 MB | +| Q2_K | [williamlcn-3400_mcq-Q2_K.gguf](https://huggingface.co/featherless-ai-quants/williamlcn-3400_mcq-GGUF/blob/main/williamlcn-3400_mcq-Q2_K.gguf) | 1172.86 MB | +| Q3_K_L | [williamlcn-3400_mcq-Q3_K_L.gguf](https://huggingface.co/featherless-ai-quants/williamlcn-3400_mcq-GGUF/blob/main/williamlcn-3400_mcq-Q3_K_L.gguf) | 1478.61 MB | +| Q3_K_M | [williamlcn-3400_mcq-Q3_K_M.gguf](https://huggingface.co/featherless-ai-quants/williamlcn-3400_mcq-GGUF/blob/main/williamlcn-3400_mcq-Q3_K_M.gguf) | 1393.96 MB | +| Q3_K_S | [williamlcn-3400_mcq-Q3_K_S.gguf](https://huggingface.co/featherless-ai-quants/williamlcn-3400_mcq-GGUF/blob/main/williamlcn-3400_mcq-Q3_K_S.gguf) | 1297.63 MB | +| Q4_K_M | [williamlcn-3400_mcq-Q4_K_M.gguf](https://huggingface.co/featherless-ai-quants/williamlcn-3400_mcq-GGUF/blob/main/williamlcn-3400_mcq-Q4_K_M.gguf) | 1629.43 MB | +| Q4_K_S | [williamlcn-3400_mcq-Q4_K_S.gguf](https://huggingface.co/featherless-ai-quants/williamlcn-3400_mcq-GGUF/blob/main/williamlcn-3400_mcq-Q4_K_S.gguf) | 1562.74 MB | +| Q5_K_M | [williamlcn-3400_mcq-Q5_K_M.gguf](https://huggingface.co/featherless-ai-quants/williamlcn-3400_mcq-GGUF/blob/main/williamlcn-3400_mcq-Q5_K_M.gguf) | 1834.18 MB | +| Q5_K_S | [williamlcn-3400_mcq-Q5_K_S.gguf](https://huggingface.co/featherless-ai-quants/williamlcn-3400_mcq-GGUF/blob/main/williamlcn-3400_mcq-Q5_K_S.gguf) | 1795.33 MB | +| Q6_K | [williamlcn-3400_mcq-Q6_K.gguf](https://huggingface.co/featherless-ai-quants/williamlcn-3400_mcq-GGUF/blob/main/williamlcn-3400_mcq-Q6_K.gguf) | 2051.73 MB | +| Q8_0 | [williamlcn-3400_mcq-Q8_0.gguf](https://huggingface.co/featherless-ai-quants/williamlcn-3400_mcq-GGUF/blob/main/williamlcn-3400_mcq-Q8_0.gguf) | 2655.50 MB | + + +--- + +## ⚡ Powered by [Featherless AI](https://featherless.ai) + +### Key Features + +- 🔥 **Instant Hosting** - Deploy any Llama model on HuggingFace instantly +- 🛠️ **Zero Infrastructure** - No server setup or maintenance required +- 📚 **Vast Compatibility** - Support for 2400+ models and counting +- 💎 **Affordable Pricing** - Starting at just $10/month + +--- + +**Links:** +[Get Started](https://featherless.ai) | [Documentation](https://featherless.ai/docs) | [Models](https://featherless.ai/models) \ No newline at end of file diff --git a/featherless-quants.png b/featherless-quants.png new file mode 100644 index 0000000..3e0eab9 --- /dev/null +++ b/featherless-quants.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e1b4d66c8306c7b0614089381fdf86ea4efb02dffb78d22767a084cb8b88d6b +size 1614532 diff --git a/williamlcn-3400_mcq-IQ4_XS.gguf b/williamlcn-3400_mcq-IQ4_XS.gguf new file mode 100644 index 0000000..5e8e2f8 --- /dev/null +++ b/williamlcn-3400_mcq-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:208ee9b6997e01c5454c71dbc902c747485f47cea7280e4354b1b54f254a71d8 +size 1576204288 diff --git a/williamlcn-3400_mcq-Q2_K.gguf b/williamlcn-3400_mcq-Q2_K.gguf new file mode 100644 index 0000000..3eac9aa --- /dev/null +++ b/williamlcn-3400_mcq-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e72ba231fd2d4b2c84b01ea905a4f7389500e148d20e1cc6a1b27f773dd1c2d +size 1229830144 diff --git a/williamlcn-3400_mcq-Q3_K_L.gguf b/williamlcn-3400_mcq-Q3_K_L.gguf new file mode 100644 index 0000000..d7f90a4 --- /dev/null +++ b/williamlcn-3400_mcq-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6be5371829ec313af97f1c8898d1147d8a08c2d26eeca005ba6bbdac025c4d93 +size 1550436352 diff --git a/williamlcn-3400_mcq-Q3_K_M.gguf b/williamlcn-3400_mcq-Q3_K_M.gguf new file mode 100644 index 0000000..1d9e7d4 --- /dev/null +++ b/williamlcn-3400_mcq-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf2b5293b87d2871b3f239b40efcf6d6c8fb230d74723fe14a5f18d66ef3c929 +size 1461667840 diff --git a/williamlcn-3400_mcq-Q3_K_S.gguf b/williamlcn-3400_mcq-Q3_K_S.gguf new file mode 100644 index 0000000..43c4739 --- /dev/null +++ b/williamlcn-3400_mcq-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a977b58fe26ae4ff627c652074e38b0742606119d05de2d5ccc953d4d76bbec7 +size 1360660480 diff --git a/williamlcn-3400_mcq-Q4_K_M.gguf b/williamlcn-3400_mcq-Q4_K_M.gguf new file mode 100644 index 0000000..4e8dfb4 --- /dev/null +++ b/williamlcn-3400_mcq-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b7ba91f0e52d249b9de74fa8fe87c40a99a6e1bf6e05d56276254135b5a5bfe +size 1708582912 diff --git a/williamlcn-3400_mcq-Q4_K_S.gguf b/williamlcn-3400_mcq-Q4_K_S.gguf new file mode 100644 index 0000000..54caab1 --- /dev/null +++ b/williamlcn-3400_mcq-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26d3eedcee96db4ecdcc00d5fc9be79a96bf965f124f55ef77ec4dda938f62e7 +size 1638651904 diff --git a/williamlcn-3400_mcq-Q5_K_M.gguf b/williamlcn-3400_mcq-Q5_K_M.gguf new file mode 100644 index 0000000..5f3fa26 --- /dev/null +++ b/williamlcn-3400_mcq-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8cdebbf3930dda3e537f3353f61210d9b4839c51cb73de8c432d94739e2e366a +size 1923278848 diff --git a/williamlcn-3400_mcq-Q5_K_S.gguf b/williamlcn-3400_mcq-Q5_K_S.gguf new file mode 100644 index 0000000..1a74b29 --- /dev/null +++ b/williamlcn-3400_mcq-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a9bca4b1bfca15fba4771b65cbc0d8dd959966ac26ea8efee25ce7b23e0c27d +size 1882544128 diff --git a/williamlcn-3400_mcq-Q6_K.gguf b/williamlcn-3400_mcq-Q6_K.gguf new file mode 100644 index 0000000..ad8bbef --- /dev/null +++ b/williamlcn-3400_mcq-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0dab024abae14ea7122829ef6095a865c71e9032274ca447d4d36ecb63bfa8e +size 2151393280 diff --git a/williamlcn-3400_mcq-Q8_0.gguf b/williamlcn-3400_mcq-Q8_0.gguf new file mode 100644 index 0000000..94f0384 --- /dev/null +++ b/williamlcn-3400_mcq-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:226485a3f9cb92e2570fc6b7b2959010d0faea0aab9358c9c4e149509d0ffe12 +size 2784495616