commit 59c547aced3d818550c98eb4071e8493cb9ec4ce Author: ModelHub XC Date: Sat May 9 14:34:37 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..938b12e --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +featherless-quants.png filter=lfs diff=lfs merge=lfs -text +magnifi-qwen2-rephrase-4bit-10-epochs-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +magnifi-qwen2-rephrase-4bit-10-epochs-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +magnifi-qwen2-rephrase-4bit-10-epochs-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +magnifi-qwen2-rephrase-4bit-10-epochs-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..8edb874 --- /dev/null +++ b/README.md @@ -0,0 +1,47 @@ +--- +base_model: magnifi/qwen2-rephrase-4bit-10-epochs +pipeline_tag: text-generation +quantized_by: featherless-ai-quants +--- + +# magnifi/qwen2-rephrase-4bit-10-epochs GGUF Quantizations 🚀 + +![Featherless AI Quants](./featherless-quants.png) + +*Optimized GGUF quantization files for enhanced model performance* + +> Powered by [Featherless AI](https://featherless.ai) - run any model you'd like for a simple small fee. +--- + +## Available Quantizations 📊 + +| Quantization Type | File | Size | +|-------------------|------|------| +| IQ4_XS | [magnifi-qwen2-rephrase-4bit-10-epochs-IQ4_XS.gguf](https://huggingface.co/featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF/blob/main/magnifi-qwen2-rephrase-4bit-10-epochs-IQ4_XS.gguf) | 335.16 MB | +| Q2_K | [magnifi-qwen2-rephrase-4bit-10-epochs-Q2_K.gguf](https://huggingface.co/featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF/blob/main/magnifi-qwen2-rephrase-4bit-10-epochs-Q2_K.gguf) | 322.92 MB | +| Q3_K_L | [magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_L.gguf](https://huggingface.co/featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF/blob/main/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_L.gguf) | 352.25 MB | +| Q3_K_M | [magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_M.gguf](https://huggingface.co/featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF/blob/main/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_M.gguf) | 339.00 MB | +| Q3_K_S | [magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_S.gguf](https://huggingface.co/featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF/blob/main/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_S.gguf) | 322.59 MB | +| Q4_K_M | [magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_M.gguf](https://huggingface.co/featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF/blob/main/magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_M.gguf) | 379.38 MB | +| Q4_K_S | [magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_S.gguf](https://huggingface.co/featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF/blob/main/magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_S.gguf) | 367.61 MB | +| Q5_K_M | [magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_M.gguf](https://huggingface.co/featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF/blob/main/magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_M.gguf) | 400.62 MB | +| Q5_K_S | [magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_S.gguf](https://huggingface.co/featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF/blob/main/magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_S.gguf) | 393.59 MB | +| Q6_K | [magnifi-qwen2-rephrase-4bit-10-epochs-Q6_K.gguf](https://huggingface.co/featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF/blob/main/magnifi-qwen2-rephrase-4bit-10-epochs-Q6_K.gguf) | 482.31 MB | +| Q8_0 | [magnifi-qwen2-rephrase-4bit-10-epochs-Q8_0.gguf](https://huggingface.co/featherless-ai-quants/magnifi-qwen2-rephrase-4bit-10-epochs-GGUF/blob/main/magnifi-qwen2-rephrase-4bit-10-epochs-Q8_0.gguf) | 506.46 MB | + + +--- + +## ⚡ Powered by [Featherless AI](https://featherless.ai) + +### Key Features + +- 🔥 **Instant Hosting** - Deploy any Llama model on HuggingFace instantly +- 🛠️ **Zero Infrastructure** - No server setup or maintenance required +- 📚 **Vast Compatibility** - Support for 2400+ models and counting +- 💎 **Affordable Pricing** - Starting at just $10/month + +--- + +**Links:** +[Get Started](https://featherless.ai) | [Documentation](https://featherless.ai/docs) | [Models](https://featherless.ai/models) \ No newline at end of file diff --git a/featherless-quants.png b/featherless-quants.png new file mode 100644 index 0000000..3e0eab9 --- /dev/null +++ b/featherless-quants.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e1b4d66c8306c7b0614089381fdf86ea4efb02dffb78d22767a084cb8b88d6b +size 1614532 diff --git a/magnifi-qwen2-rephrase-4bit-10-epochs-IQ4_XS.gguf b/magnifi-qwen2-rephrase-4bit-10-epochs-IQ4_XS.gguf new file mode 100644 index 0000000..bef398d --- /dev/null +++ b/magnifi-qwen2-rephrase-4bit-10-epochs-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e47684dc31e90dbe0393af0ada936825bcd8273b015ba1ec4860e93a5d3d02b9 +size 351442944 diff --git a/magnifi-qwen2-rephrase-4bit-10-epochs-Q2_K.gguf b/magnifi-qwen2-rephrase-4bit-10-epochs-Q2_K.gguf new file mode 100644 index 0000000..20470b9 --- /dev/null +++ b/magnifi-qwen2-rephrase-4bit-10-epochs-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ddf12aaa49454bff5ded70af90a87107c56aa9baf0722a1d08ec3eb218a7f960 +size 338605056 diff --git a/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_L.gguf b/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_L.gguf new file mode 100644 index 0000000..891d11f --- /dev/null +++ b/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8b2af0b510e401dc12473d4a8a90313aecd5b8b97760871c3358781f16f1a0e +size 369355776 diff --git a/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_M.gguf b/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_M.gguf new file mode 100644 index 0000000..d055bd7 --- /dev/null +++ b/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81c3af11f0ac5e64c172eabec90387d82b1f35d857be9c50d100c961960f8874 +size 355464192 diff --git a/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_S.gguf b/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_S.gguf new file mode 100644 index 0000000..41cf85b --- /dev/null +++ b/magnifi-qwen2-rephrase-4bit-10-epochs-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8adb5d0780eb492c43c97d985a2b1007736351dfa2ea179200650a1a9b7a2e61 +size 338260992 diff --git a/magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_M.gguf b/magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_M.gguf new file mode 100644 index 0000000..a2c5240 --- /dev/null +++ b/magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1399705b16b8a9cc2069e2775cd8706a6e2f3d428c20f0181d999d0329e4f249 +size 397805568 diff --git a/magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_S.gguf b/magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_S.gguf new file mode 100644 index 0000000..0d9d38f --- /dev/null +++ b/magnifi-qwen2-rephrase-4bit-10-epochs-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3f283805b2e1ea55ab26e2794a9b7c6485b720e53401a3d4914ab9fa87f73a0 +size 385469440 diff --git a/magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_M.gguf b/magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_M.gguf new file mode 100644 index 0000000..b0ff025 --- /dev/null +++ b/magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1ec0bd9898a78eff82f6101e812f4a165e16eb0573b9b1c187b7d2c7cf7bd38 +size 420083712 diff --git a/magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_S.gguf b/magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_S.gguf new file mode 100644 index 0000000..ee0a293 --- /dev/null +++ b/magnifi-qwen2-rephrase-4bit-10-epochs-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e21c324a8651c57c9353a215c7a7fcb4d58720a56c2a6666c33b80301f211fb5 +size 412707840 diff --git a/magnifi-qwen2-rephrase-4bit-10-epochs-Q6_K.gguf b/magnifi-qwen2-rephrase-4bit-10-epochs-Q6_K.gguf new file mode 100644 index 0000000..9ddc256 --- /dev/null +++ b/magnifi-qwen2-rephrase-4bit-10-epochs-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ab6d4e2af386062a8bb508c6996a2c783c8df039bdc0c80869c9d7e4be1f89c +size 505734144 diff --git a/magnifi-qwen2-rephrase-4bit-10-epochs-Q8_0.gguf b/magnifi-qwen2-rephrase-4bit-10-epochs-Q8_0.gguf new file mode 100644 index 0000000..cd7b805 --- /dev/null +++ b/magnifi-qwen2-rephrase-4bit-10-epochs-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3417757d722d07049a4b14055fa591334e05d7a859a2e441d5c6832dd3f93e1d +size 531065856