commit e4c21c8846a896ca804db2fa4fa30ac998ffac1f Author: ModelHub XC Date: Sat May 9 14:52:36 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: featherless-ai-quants/google-gemma-1.1-2b-it-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..eb259e3 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +featherless-quants.png filter=lfs diff=lfs merge=lfs -text +google-gemma-1.1-2b-it-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-1.1-2b-it-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-1.1-2b-it-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-1.1-2b-it-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-1.1-2b-it-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-1.1-2b-it-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-1.1-2b-it-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-1.1-2b-it-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-1.1-2b-it-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-1.1-2b-it-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +google-gemma-1.1-2b-it-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..3c31c1d --- /dev/null +++ b/README.md @@ -0,0 +1,47 @@ +--- +base_model: google/gemma-1.1-2b-it +pipeline_tag: text-generation +quantized_by: featherless-ai-quants +--- + +# google/gemma-1.1-2b-it GGUF Quantizations 🚀 + +![Featherless AI Quants](./featherless-quants.png) + +*Optimized GGUF quantization files for enhanced model performance* + +> Powered by [Featherless AI](https://featherless.ai) - run any model you'd like for a simple small fee. +--- + +## Available Quantizations 📊 + +| Quantization Type | File | Size | +|-------------------|------|------| +| IQ4_XS | [google-gemma-1.1-2b-it-IQ4_XS.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-1.1-2b-it-GGUF/blob/main/google-gemma-1.1-2b-it-IQ4_XS.gguf) | 1431.67 MB | +| Q2_K | [google-gemma-1.1-2b-it-Q2_K.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-1.1-2b-it-GGUF/blob/main/google-gemma-1.1-2b-it-Q2_K.gguf) | 1104.28 MB | +| Q3_K_L | [google-gemma-1.1-2b-it-Q3_K_L.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-1.1-2b-it-GGUF/blob/main/google-gemma-1.1-2b-it-Q3_K_L.gguf) | 1397.70 MB | +| Q3_K_M | [google-gemma-1.1-2b-it-Q3_K_M.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-1.1-2b-it-GGUF/blob/main/google-gemma-1.1-2b-it-Q3_K_M.gguf) | 1319.70 MB | +| Q3_K_S | [google-gemma-1.1-2b-it-Q3_K_S.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-1.1-2b-it-GGUF/blob/main/google-gemma-1.1-2b-it-Q3_K_S.gguf) | 1228.31 MB | +| Q4_K_M | [google-gemma-1.1-2b-it-Q4_K_M.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-1.1-2b-it-GGUF/blob/main/google-gemma-1.1-2b-it-Q4_K_M.gguf) | 1554.74 MB | +| Q4_K_S | [google-gemma-1.1-2b-it-Q4_K_S.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-1.1-2b-it-GGUF/blob/main/google-gemma-1.1-2b-it-Q4_K_S.gguf) | 1487.58 MB | +| Q5_K_M | [google-gemma-1.1-2b-it-Q5_K_M.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-1.1-2b-it-GGUF/blob/main/google-gemma-1.1-2b-it-Q5_K_M.gguf) | 1754.43 MB | +| Q5_K_S | [google-gemma-1.1-2b-it-Q5_K_S.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-1.1-2b-it-GGUF/blob/main/google-gemma-1.1-2b-it-Q5_K_S.gguf) | 1715.58 MB | +| Q6_K | [google-gemma-1.1-2b-it-Q6_K.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-1.1-2b-it-GGUF/blob/main/google-gemma-1.1-2b-it-Q6_K.gguf) | 1966.60 MB | +| Q8_0 | [google-gemma-1.1-2b-it-Q8_0.gguf](https://huggingface.co/featherless-ai-quants/google-gemma-1.1-2b-it-GGUF/blob/main/google-gemma-1.1-2b-it-Q8_0.gguf) | 2545.42 MB | + + +--- + +## ⚡ Powered by [Featherless AI](https://featherless.ai) + +### Key Features + +- 🔥 **Instant Hosting** - Deploy any Llama model on HuggingFace instantly +- 🛠️ **Zero Infrastructure** - No server setup or maintenance required +- 📚 **Vast Compatibility** - Support for 2400+ models and counting +- 💎 **Affordable Pricing** - Starting at just $10/month + +--- + +**Links:** +[Get Started](https://featherless.ai) | [Documentation](https://featherless.ai/docs) | [Models](https://featherless.ai/models) \ No newline at end of file diff --git a/featherless-quants.png b/featherless-quants.png new file mode 100644 index 0000000..3e0eab9 --- /dev/null +++ b/featherless-quants.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e1b4d66c8306c7b0614089381fdf86ea4efb02dffb78d22767a084cb8b88d6b +size 1614532 diff --git a/google-gemma-1.1-2b-it-IQ4_XS.gguf b/google-gemma-1.1-2b-it-IQ4_XS.gguf new file mode 100644 index 0000000..4fa4c2f --- /dev/null +++ b/google-gemma-1.1-2b-it-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:efa1cb57d7a6f9f6e12902b9e0d633cedd7855c81db0fbc7533757052cb8c49b +size 1501218880 diff --git a/google-gemma-1.1-2b-it-Q2_K.gguf b/google-gemma-1.1-2b-it-Q2_K.gguf new file mode 100644 index 0000000..80d387b --- /dev/null +++ b/google-gemma-1.1-2b-it-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff27dc49413468e465528e36e10003e08d9cacae710e2cad1210d96e9533dbe5 +size 1157924928 diff --git a/google-gemma-1.1-2b-it-Q3_K_L.gguf b/google-gemma-1.1-2b-it-Q3_K_L.gguf new file mode 100644 index 0000000..14dc0fc --- /dev/null +++ b/google-gemma-1.1-2b-it-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b5c598adc014925f151baf05062b1fdc856073364f910a1b8431cdb6aef766 +size 1465591872 diff --git a/google-gemma-1.1-2b-it-Q3_K_M.gguf b/google-gemma-1.1-2b-it-Q3_K_M.gguf new file mode 100644 index 0000000..62ac56c --- /dev/null +++ b/google-gemma-1.1-2b-it-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4eba3aff291faa00005e4560f9bb19ddbc20183e6ec849a493fdf6976541b25e +size 1383802944 diff --git a/google-gemma-1.1-2b-it-Q3_K_S.gguf b/google-gemma-1.1-2b-it-Q3_K_S.gguf new file mode 100644 index 0000000..4544040 --- /dev/null +++ b/google-gemma-1.1-2b-it-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:770678907daf35dfb2ea5ce6969e172215b7ee21c4f692cef16f24502d377e31 +size 1287981120 diff --git a/google-gemma-1.1-2b-it-Q4_K_M.gguf b/google-gemma-1.1-2b-it-Q4_K_M.gguf new file mode 100644 index 0000000..95a49bf --- /dev/null +++ b/google-gemma-1.1-2b-it-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5fad6a7a16491ebb0a44cc9455fc78def4812986b6935b1c0c6faa9b62f3a009 +size 1630263360 diff --git a/google-gemma-1.1-2b-it-Q4_K_S.gguf b/google-gemma-1.1-2b-it-Q4_K_S.gguf new file mode 100644 index 0000000..e4584ca --- /dev/null +++ b/google-gemma-1.1-2b-it-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f201a23e8fbc2f7248bd10db3ab17c07c06b182556439766df2bb416fffc85aa +size 1559840832 diff --git a/google-gemma-1.1-2b-it-Q5_K_M.gguf b/google-gemma-1.1-2b-it-Q5_K_M.gguf new file mode 100644 index 0000000..1169d6d --- /dev/null +++ b/google-gemma-1.1-2b-it-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1125a5f0c0447ad8443b09c878df577485ebde051c0c2e3944d97e433e89a896 +size 1839650880 diff --git a/google-gemma-1.1-2b-it-Q5_K_S.gguf b/google-gemma-1.1-2b-it-Q5_K_S.gguf new file mode 100644 index 0000000..4875e32 --- /dev/null +++ b/google-gemma-1.1-2b-it-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61367ec6a0d20dcaff28544cf05d4c0043633508aefacdc46ffb9ac0a26c2635 +size 1798916160 diff --git a/google-gemma-1.1-2b-it-Q6_K.gguf b/google-gemma-1.1-2b-it-Q6_K.gguf new file mode 100644 index 0000000..c5b3fde --- /dev/null +++ b/google-gemma-1.1-2b-it-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0b6c0eb28a6e9eeb9f63f9728890aeb12d729fa231ec3d42f3f5b4ca7320705 +size 2062125120 diff --git a/google-gemma-1.1-2b-it-Q8_0.gguf b/google-gemma-1.1-2b-it-Q8_0.gguf new file mode 100644 index 0000000..bd5cf72 --- /dev/null +++ b/google-gemma-1.1-2b-it-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c9f347eeb06106b4d2da4bef217744a3acc4845925bdc5a760e1eb226fd0e91 +size 2669070400