commit 9225e253e850052398a13e2cac704622e946d4b0 Author: ModelHub XC Date: Tue Apr 21 18:11:50 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: featherless-ai-quants/google-gemma-2-2b-it-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..a25e890 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +featherless-quants.png filter=lfs diff=lfs merge=lfs -text +google-gemma-2-2b-it-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-2-2b-it-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-2-2b-it-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-2-2b-it-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-2-2b-it-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-2-2b-it-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-2-2b-it-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-2-2b-it-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-2-2b-it-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-2-2b-it-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-2-2b-it-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..e2137e5 --- /dev/null +++ b/README.md @@ -0,0 +1,47 @@ +--- +base_model: google/gemma-2-2b-it +pipeline_tag: text-generation +quantized_by: featherless-ai-quants +--- + +# google/gemma-2-2b-it GGUF Quantizations 🚀 + +![Featherless AI Quants](./featherless-quants.png) + +*Optimized GGUF quantization files for enhanced model performance* + +> Powered by [Featherless AI](https://featherless.ai) - run any model you'd like for a simple small fee. +--- + +## Available Quantizations 📊 + +| Quantization Type | File | Size | +|-------------------|------|------| +| IQ4_XS | [google-gemma-2-2b-it-IQ4_XS.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-2-2b-it-GGUF/blob/main/google-gemma-2-2b-it-IQ4_XS.gguf) | 1503.19 MB | +| Q2_K | [google-gemma-2-2b-it-Q2_K.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-2-2b-it-GGUF/blob/main/google-gemma-2-2b-it-Q2_K.gguf) | 1172.86 MB | +| Q3_K_L | [google-gemma-2-2b-it-Q3_K_L.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-2-2b-it-GGUF/blob/main/google-gemma-2-2b-it-Q3_K_L.gguf) | 1478.61 MB | +| Q3_K_M | [google-gemma-2-2b-it-Q3_K_M.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-2-2b-it-GGUF/blob/main/google-gemma-2-2b-it-Q3_K_M.gguf) | 1393.95 MB | +| Q3_K_S | [google-gemma-2-2b-it-Q3_K_S.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-2-2b-it-GGUF/blob/main/google-gemma-2-2b-it-Q3_K_S.gguf) | 1297.63 MB | +| Q4_K_M | [google-gemma-2-2b-it-Q4_K_M.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-2-2b-it-GGUF/blob/main/google-gemma-2-2b-it-Q4_K_M.gguf) | 1629.43 MB | +| Q4_K_S | [google-gemma-2-2b-it-Q4_K_S.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-2-2b-it-GGUF/blob/main/google-gemma-2-2b-it-Q4_K_S.gguf) | 1562.74 MB | +| Q5_K_M | [google-gemma-2-2b-it-Q5_K_M.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-2-2b-it-GGUF/blob/main/google-gemma-2-2b-it-Q5_K_M.gguf) | 1834.18 MB | +| Q5_K_S | [google-gemma-2-2b-it-Q5_K_S.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-2-2b-it-GGUF/blob/main/google-gemma-2-2b-it-Q5_K_S.gguf) | 1795.33 MB | +| Q6_K | [google-gemma-2-2b-it-Q6_K.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-2-2b-it-GGUF/blob/main/google-gemma-2-2b-it-Q6_K.gguf) | 2051.73 MB | +| Q8_0 | [google-gemma-2-2b-it-Q8_0.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-2-2b-it-GGUF/blob/main/google-gemma-2-2b-it-Q8_0.gguf) | 2655.50 MB | + + +--- + +## ⚡ Powered by [Featherless AI](https://featherless.ai) + +### Key Features + +- 🔥 **Instant Hosting** - Deploy any Llama model on HuggingFace instantly +- 🛠️ **Zero Infrastructure** - No server setup or maintenance required +- 📚 **Vast Compatibility** - Support for 2400+ models and counting +- 💎 **Affordable Pricing** - Starting at just $10/month + +--- + +**Links:** +[Get Started](https://featherless.ai) | [Documentation](https://featherless.ai/docs) | [Models](https://featherless.ai/models) \ No newline at end of file diff --git a/featherless-quants.png b/featherless-quants.png new file mode 100644 index 0000000..3e0eab9 --- /dev/null +++ b/featherless-quants.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e1b4d66c8306c7b0614089381fdf86ea4efb02dffb78d22767a084cb8b88d6b +size 1614532 diff --git a/google-gemma-2-2b-it-IQ4_XS.gguf b/google-gemma-2-2b-it-IQ4_XS.gguf new file mode 100644 index 0000000..bd38746 --- /dev/null +++ b/google-gemma-2-2b-it-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f7fc81ef633e0436d5def99ce90c531682f5b331ace4842ee9e50a979c1ca7d +size 1576204128 diff --git a/google-gemma-2-2b-it-Q2_K.gguf b/google-gemma-2-2b-it-Q2_K.gguf new file mode 100644 index 0000000..733c04b --- /dev/null +++ b/google-gemma-2-2b-it-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a603f6bbdf4472532aba4a3033a38f6dc041ec4ee7a1bbbd42bb7f1bc964234a +size 1229829984 diff --git a/google-gemma-2-2b-it-Q3_K_L.gguf b/google-gemma-2-2b-it-Q3_K_L.gguf new file mode 100644 index 0000000..5fdeab7 --- /dev/null +++ b/google-gemma-2-2b-it-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:228a2f872e5e9d7076d75a81fb71617bc1147084836be7a18633c04dd7c3d9a8 +size 1550436192 diff --git a/google-gemma-2-2b-it-Q3_K_M.gguf b/google-gemma-2-2b-it-Q3_K_M.gguf new file mode 100644 index 0000000..ddc3e57 --- /dev/null +++ b/google-gemma-2-2b-it-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:368a0b05018feec39c00ecec3db2ecea8691fe67e8f02b69f50eba6b471a51b2 +size 1461667680 diff --git a/google-gemma-2-2b-it-Q3_K_S.gguf b/google-gemma-2-2b-it-Q3_K_S.gguf new file mode 100644 index 0000000..fcee7fb --- /dev/null +++ b/google-gemma-2-2b-it-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3da090c52920d60c16683f7bf3ed8f437f2da82794482c376a6544697bb8bf98 +size 1360660320 diff --git a/google-gemma-2-2b-it-Q4_K_M.gguf b/google-gemma-2-2b-it-Q4_K_M.gguf new file mode 100644 index 0000000..fcac559 --- /dev/null +++ b/google-gemma-2-2b-it-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5cee79eccf661269abecf19c4ca4728825678596046a8be6dc22df3d3188832 +size 1708582752 diff --git a/google-gemma-2-2b-it-Q4_K_S.gguf b/google-gemma-2-2b-it-Q4_K_S.gguf new file mode 100644 index 0000000..b5516b3 --- /dev/null +++ b/google-gemma-2-2b-it-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fcf26be2c4b732d88e0198ace456dd8119c81cf122a978ec1d8a0cd689091b8 +size 1638651744 diff --git a/google-gemma-2-2b-it-Q5_K_M.gguf b/google-gemma-2-2b-it-Q5_K_M.gguf new file mode 100644 index 0000000..8240dfb --- /dev/null +++ b/google-gemma-2-2b-it-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbc6d5fde8a5ac5d41443035e81cf8bd6f5f11876b7abd82edac3dbc3f52c393 +size 1923278688 diff --git a/google-gemma-2-2b-it-Q5_K_S.gguf b/google-gemma-2-2b-it-Q5_K_S.gguf new file mode 100644 index 0000000..807cf81 --- /dev/null +++ b/google-gemma-2-2b-it-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1dbcaad75cfa7d2012cb481c6507b970d3ab19370f847d769b57bcb29c9f772c +size 1882543968 diff --git a/google-gemma-2-2b-it-Q6_K.gguf b/google-gemma-2-2b-it-Q6_K.gguf new file mode 100644 index 0000000..0201c03 --- /dev/null +++ b/google-gemma-2-2b-it-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9f241bb32eae37e55a15efc50d0a66986a55cccf8f107aafd4b762cd39e86a34 +size 2151393120 diff --git a/google-gemma-2-2b-it-Q8_0.gguf b/google-gemma-2-2b-it-Q8_0.gguf new file mode 100644 index 0000000..ceb9440 --- /dev/null +++ b/google-gemma-2-2b-it-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c97f22c66a86aee064c55acddd46562a6769afe27ecbb87622f5ef0b2f12ea91 +size 2784495456