commit 82a93cd64d5eafa41b81a1273219c96617944c86 Author: ModelHub XC Date: Sat May 2 20:32:07 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/Aura-7b-i1-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..f26afa5 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Aura-7b.imatrix.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +Aura-7b.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Aura-7b.i1-IQ1_M.gguf b/Aura-7b.i1-IQ1_M.gguf new file mode 100644 index 0000000..1cb6dbc --- /dev/null +++ b/Aura-7b.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f30cda9b530dc1893562379da498b9da37166da7ca96da1ee0e399323f13720 +size 2042197056 diff --git a/Aura-7b.i1-IQ1_S.gguf b/Aura-7b.i1-IQ1_S.gguf new file mode 100644 index 0000000..67f84d7 --- /dev/null +++ b/Aura-7b.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c4cfe323d60a77abf8e331f6dbff1e99856a2c7189db2f8fbef583876787b5b +size 1903668288 diff --git a/Aura-7b.i1-IQ2_M.gguf b/Aura-7b.i1-IQ2_M.gguf new file mode 100644 index 0000000..186eafa --- /dev/null +++ b/Aura-7b.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6455d25cfb476d3b1337071b0fd0b4238209a857d6d3af655721a447a65f0de +size 2780343360 diff --git a/Aura-7b.i1-IQ2_S.gguf b/Aura-7b.i1-IQ2_S.gguf new file mode 100644 index 0000000..00634f0 --- /dev/null +++ b/Aura-7b.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df43ef062b51942ab57506173d03ef7baf059ede183a14590121e81dab83fcb7 +size 2595638336 diff --git a/Aura-7b.i1-IQ2_XS.gguf b/Aura-7b.i1-IQ2_XS.gguf new file mode 100644 index 0000000..63df7d9 --- /dev/null +++ b/Aura-7b.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01de2c3132410508ac298fda9ec87a20d4952383e075adfb810b3e1ba7c00fba +size 2469022784 diff --git a/Aura-7b.i1-IQ2_XXS.gguf b/Aura-7b.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..55fe44a --- /dev/null +++ b/Aura-7b.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b180847b2fb2f22f4b1f90d54fad16063f85d6756d52f236523e9378f8fcfdd +size 2273078336 diff --git a/Aura-7b.i1-IQ3_M.gguf b/Aura-7b.i1-IQ3_M.gguf new file mode 100644 index 0000000..71d3b7c --- /dev/null +++ b/Aura-7b.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eaad19cb95ba1564cb065f2f1b2ce7002eb4d20467ea58e71efcfc495a0da68b +size 3574012992 diff --git a/Aura-7b.i1-IQ3_S.gguf b/Aura-7b.i1-IQ3_S.gguf new file mode 100644 index 0000000..54d13ba --- /dev/null +++ b/Aura-7b.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e4b0ff56fab2c731ec2dd9dd8a8a7331bee3a7ddd5c453e1df1f7421785ca43 +size 3499193408 diff --git a/Aura-7b.i1-IQ3_XS.gguf b/Aura-7b.i1-IQ3_XS.gguf new file mode 100644 index 0000000..e740275 --- /dev/null +++ b/Aura-7b.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8939e6b39175f42e7e994028fd3c5f17d55a5fbf9b3eca361f3d0e2ecfc07e4 +size 3346256960 diff --git a/Aura-7b.i1-IQ3_XXS.gguf b/Aura-7b.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..3c00381 --- /dev/null +++ b/Aura-7b.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:292bd98470121d66fd00e60fa7bf6cffbaef964aae489c166cc65c2962b6bbc5 +size 3114515520 diff --git a/Aura-7b.i1-IQ4_NL.gguf b/Aura-7b.i1-IQ4_NL.gguf new file mode 100644 index 0000000..10cc5c0 --- /dev/null +++ b/Aura-7b.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a84abad42b8451fd239fd2326c1a61102a132eb88368c57f2c3027896b7b275c +size 4437814336 diff --git a/Aura-7b.i1-IQ4_XS.gguf b/Aura-7b.i1-IQ4_XS.gguf new file mode 100644 index 0000000..f60cfe6 --- /dev/null +++ b/Aura-7b.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4858759c1e5f10bc185aef0288b6b75303842296deae1f83c6d7f3418d74ce7 +size 4218473536 diff --git a/Aura-7b.i1-Q2_K.gguf b/Aura-7b.i1-Q2_K.gguf new file mode 100644 index 0000000..6f710e1 --- /dev/null +++ b/Aura-7b.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3891ef958089d06652733eead3bdff36bd5551c0a3a4c3a13bb50636f3605c07 +size 3015941184 diff --git a/Aura-7b.i1-Q2_K_S.gguf b/Aura-7b.i1-Q2_K_S.gguf new file mode 100644 index 0000000..d56e312 --- /dev/null +++ b/Aura-7b.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03590739b70de50d5f8360e848c90aa4d56e587feb4b7c90fd38b48b3c72636f +size 2834074688 diff --git a/Aura-7b.i1-Q3_K_L.gguf b/Aura-7b.i1-Q3_K_L.gguf new file mode 100644 index 0000000..53e4827 --- /dev/null +++ b/Aura-7b.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1eb14e9d5785807089066dcaa526416c5313a26c8f3db7bf2ab488885bef0b8 +size 4088460352 diff --git a/Aura-7b.i1-Q3_K_M.gguf b/Aura-7b.i1-Q3_K_M.gguf new file mode 100644 index 0000000..42fde66 --- /dev/null +++ b/Aura-7b.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e7f5fa55e9b0e644a9c13560759a8708242759c228989c0f17225081d0581dd +size 3808392256 diff --git a/Aura-7b.i1-Q3_K_S.gguf b/Aura-7b.i1-Q3_K_S.gguf new file mode 100644 index 0000000..cec1136 --- /dev/null +++ b/Aura-7b.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd64ec07229364dc84a5dd7d96197f5647de325c92bf39d9b246815e80f08a42 +size 3492369472 diff --git a/Aura-7b.i1-Q4_0.gguf b/Aura-7b.i1-Q4_0.gguf new file mode 100644 index 0000000..fb1f561 --- /dev/null +++ b/Aura-7b.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc7bf6c857c2c64ac83d71a5c1864052f3f6682199fec5d422cb63e106db498d +size 4444122176 diff --git a/Aura-7b.i1-Q4_1.gguf b/Aura-7b.i1-Q4_1.gguf new file mode 100644 index 0000000..f920b28 --- /dev/null +++ b/Aura-7b.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:232f8a9fee169054e87f31b56c04c2becf1d4f1ebbb794caa737b1bd48e60f31 +size 4873284672 diff --git a/Aura-7b.i1-Q4_K_M.gguf b/Aura-7b.i1-Q4_K_M.gguf new file mode 100644 index 0000000..d5c5683 --- /dev/null +++ b/Aura-7b.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad673572699bd424b61c8d5c70ce0c6f566b90724048b5613782ceef4389f5e1 +size 4683074624 diff --git a/Aura-7b.i1-Q4_K_S.gguf b/Aura-7b.i1-Q4_K_S.gguf new file mode 100644 index 0000000..9f69a5b --- /dev/null +++ b/Aura-7b.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e90279815b67d1a7a334fdf7434672eda7d06cc1a399ab5708dc2c931564a8d +size 4457770048 diff --git a/Aura-7b.i1-Q5_K_M.gguf b/Aura-7b.i1-Q5_K_M.gguf new file mode 100644 index 0000000..c5e5dbb --- /dev/null +++ b/Aura-7b.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef9a44f71b3eac52592f092e60933089ec1d4400729efaab0a38527ed664e331 +size 5444832320 diff --git a/Aura-7b.i1-Q5_K_S.gguf b/Aura-7b.i1-Q5_K_S.gguf new file mode 100644 index 0000000..58c3608 --- /dev/null +++ b/Aura-7b.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fab562120fa9d3a86e166f979f92eb99f9659317c29ef6c27f6f289348fe8235 +size 5315177536 diff --git a/Aura-7b.i1-Q6_K.gguf b/Aura-7b.i1-Q6_K.gguf new file mode 100644 index 0000000..17465f3 --- /dev/null +++ b/Aura-7b.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95d15242745186bb959782ec18bf5f1843a4f0da64089a7fcd2470f3b2824ecc +size 6254199872 diff --git a/Aura-7b.imatrix.gguf b/Aura-7b.imatrix.gguf new file mode 100644 index 0000000..638e62f --- /dev/null +++ b/Aura-7b.imatrix.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83b64fbea2c00c0e1c14845dd5a8e94ff4c62d154b8cd134036220972df8f78f +size 4560352 diff --git a/README.md b/README.md new file mode 100644 index 0000000..91b067e --- /dev/null +++ b/README.md @@ -0,0 +1,96 @@ +--- +base_model: Featherlabs/Aura-7b +datasets: +- Featherlabs/featherlabs_agentic_v1 +language: +- en +library_name: transformers +license: apache-2.0 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- qwen2 +- unsloth +- trl +- sft +- agentic +- function-calling +- tool-use +- featherlabs +--- +## About + + + + + + + + + +weighted/imatrix quants of https://huggingface.co/Featherlabs/Aura-7b + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Aura-7b-i1-GGUF).*** + +static quants are available at https://huggingface.co/mradermacher/Aura-7b-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.imatrix.gguf) | imatrix | 0.1 | imatrix file (for creating your own quants) | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ1_S.gguf) | i1-IQ1_S | 2.0 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ1_M.gguf) | i1-IQ1_M | 2.1 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 2.4 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ2_XS.gguf) | i1-IQ2_XS | 2.6 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ2_S.gguf) | i1-IQ2_S | 2.7 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ2_M.gguf) | i1-IQ2_M | 2.9 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q2_K_S.gguf) | i1-Q2_K_S | 2.9 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q2_K.gguf) | i1-Q2_K | 3.1 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 3.2 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ3_XS.gguf) | i1-IQ3_XS | 3.4 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q3_K_S.gguf) | i1-Q3_K_S | 3.6 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ3_S.gguf) | i1-IQ3_S | 3.6 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ3_M.gguf) | i1-IQ3_M | 3.7 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q3_K_M.gguf) | i1-Q3_K_M | 3.9 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q3_K_L.gguf) | i1-Q3_K_L | 4.2 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ4_XS.gguf) | i1-IQ4_XS | 4.3 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-IQ4_NL.gguf) | i1-IQ4_NL | 4.5 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q4_0.gguf) | i1-Q4_0 | 4.5 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q4_K_S.gguf) | i1-Q4_K_S | 4.6 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q4_K_M.gguf) | i1-Q4_K_M | 4.8 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q4_1.gguf) | i1-Q4_1 | 5.0 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q5_K_S.gguf) | i1-Q5_K_S | 5.4 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q5_K_M.gguf) | i1-Q5_K_M | 5.5 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-7b-i1-GGUF/resolve/main/Aura-7b.i1-Q6_K.gguf) | i1-Q6_K | 6.4 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + +