From 0bcbd4d252a2d297f99f5d6a4c25f0d9e3ec8b87 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Fri, 10 Apr 2026 12:15:53 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/Aura-4B-i1-GGUF Source: Original Platform --- .gitattributes | 58 ++++++++++++++++++++++++++++++ Aura-4B.i1-IQ1_M.gguf | 3 ++ Aura-4B.i1-IQ1_S.gguf | 3 ++ Aura-4B.i1-IQ2_M.gguf | 3 ++ Aura-4B.i1-IQ2_S.gguf | 3 ++ Aura-4B.i1-IQ2_XS.gguf | 3 ++ Aura-4B.i1-IQ2_XXS.gguf | 3 ++ Aura-4B.i1-IQ3_M.gguf | 3 ++ Aura-4B.i1-IQ3_S.gguf | 3 ++ Aura-4B.i1-IQ3_XS.gguf | 3 ++ Aura-4B.i1-IQ3_XXS.gguf | 3 ++ Aura-4B.i1-IQ4_XS.gguf | 3 ++ Aura-4B.i1-Q2_K.gguf | 3 ++ Aura-4B.i1-Q2_K_S.gguf | 3 ++ Aura-4B.i1-Q3_K_L.gguf | 3 ++ Aura-4B.i1-Q3_K_M.gguf | 3 ++ Aura-4B.i1-Q3_K_S.gguf | 3 ++ Aura-4B.i1-Q4_0.gguf | 3 ++ Aura-4B.i1-Q4_K_M.gguf | 3 ++ Aura-4B.i1-Q4_K_S.gguf | 3 ++ Aura-4B.i1-Q5_K_M.gguf | 3 ++ Aura-4B.i1-Q5_K_S.gguf | 3 ++ Aura-4B.i1-Q6_K.gguf | 3 ++ README.md | 79 +++++++++++++++++++++++++++++++++++++++++ imatrix.dat | 3 ++ 25 files changed, 206 insertions(+) create mode 100644 .gitattributes create mode 100644 Aura-4B.i1-IQ1_M.gguf create mode 100644 Aura-4B.i1-IQ1_S.gguf create mode 100644 Aura-4B.i1-IQ2_M.gguf create mode 100644 Aura-4B.i1-IQ2_S.gguf create mode 100644 Aura-4B.i1-IQ2_XS.gguf create mode 100644 Aura-4B.i1-IQ2_XXS.gguf create mode 100644 Aura-4B.i1-IQ3_M.gguf create mode 100644 Aura-4B.i1-IQ3_S.gguf create mode 100644 Aura-4B.i1-IQ3_XS.gguf create mode 100644 Aura-4B.i1-IQ3_XXS.gguf create mode 100644 Aura-4B.i1-IQ4_XS.gguf create mode 100644 Aura-4B.i1-Q2_K.gguf create mode 100644 Aura-4B.i1-Q2_K_S.gguf create mode 100644 Aura-4B.i1-Q3_K_L.gguf create mode 100644 Aura-4B.i1-Q3_K_M.gguf create mode 100644 Aura-4B.i1-Q3_K_S.gguf create mode 100644 Aura-4B.i1-Q4_0.gguf create mode 100644 Aura-4B.i1-Q4_K_M.gguf create mode 100644 Aura-4B.i1-Q4_K_S.gguf create mode 100644 Aura-4B.i1-Q5_K_M.gguf create mode 100644 Aura-4B.i1-Q5_K_S.gguf create mode 100644 Aura-4B.i1-Q6_K.gguf create mode 100644 README.md create mode 100644 imatrix.dat diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..17bbbab --- /dev/null +++ b/.gitattributes @@ -0,0 +1,58 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +imatrix.dat filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +Aura-4B.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Aura-4B.i1-IQ1_M.gguf b/Aura-4B.i1-IQ1_M.gguf new file mode 100644 index 0000000..8dcfc59 --- /dev/null +++ b/Aura-4B.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae770d38fff90af95fd6e1878d62ad5500d3b9c8a3ee180d9377d50026be937a +size 1286257344 diff --git a/Aura-4B.i1-IQ1_S.gguf b/Aura-4B.i1-IQ1_S.gguf new file mode 100644 index 0000000..bbc1944 --- /dev/null +++ b/Aura-4B.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d34a74911f677a3b8cc63b91ddad366a8a0bd192ce59cf4d6cb0e7ffd409ba5 +size 1213414080 diff --git a/Aura-4B.i1-IQ2_M.gguf b/Aura-4B.i1-IQ2_M.gguf new file mode 100644 index 0000000..2c4e318 --- /dev/null +++ b/Aura-4B.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b5f21bf9a8c43bca58e7b013900690a4bd54419ee1d1d6184dabc3570dd4eae +size 1722634944 diff --git a/Aura-4B.i1-IQ2_S.gguf b/Aura-4B.i1-IQ2_S.gguf new file mode 100644 index 0000000..bfd8e09 --- /dev/null +++ b/Aura-4B.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:994328b9e955f0153fa331f03f939efed17985cb2ae034ebe3df69da673386c4 +size 1625510592 diff --git a/Aura-4B.i1-IQ2_XS.gguf b/Aura-4B.i1-IQ2_XS.gguf new file mode 100644 index 0000000..634c7e1 --- /dev/null +++ b/Aura-4B.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05d2d8a71f28227d832299514473ea5de824687d5d4bd7321d6538401ad4b30c +size 1517370048 diff --git a/Aura-4B.i1-IQ2_XXS.gguf b/Aura-4B.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..5f43555 --- /dev/null +++ b/Aura-4B.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:611a62debb6ad03ea5f24b85f6a983ed581b57a97c49f1c6e597e2d7b18595f8 +size 1407662784 diff --git a/Aura-4B.i1-IQ3_M.gguf b/Aura-4B.i1-IQ3_M.gguf new file mode 100644 index 0000000..b4f1b7f --- /dev/null +++ b/Aura-4B.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5092eb22f07ea21a315d149d11a5471ac44052b2ccc49e24fce9d87c9cf670b7 +size 2183416512 diff --git a/Aura-4B.i1-IQ3_S.gguf b/Aura-4B.i1-IQ3_S.gguf new file mode 100644 index 0000000..523003f --- /dev/null +++ b/Aura-4B.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3ac2ea0377269a701161d456d2d6deb5c8cf5bf96df6e4c5d7b2a5f73e5c028 +size 2114898624 diff --git a/Aura-4B.i1-IQ3_XS.gguf b/Aura-4B.i1-IQ3_XS.gguf new file mode 100644 index 0000000..5c2faf9 --- /dev/null +++ b/Aura-4B.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c97772359a5fe1b4e8f7b9f521408f0a9dadf91006e882c4623632ad88e2c186 +size 2027604672 diff --git a/Aura-4B.i1-IQ3_XXS.gguf b/Aura-4B.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..bbbdbfd --- /dev/null +++ b/Aura-4B.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66720da4ccbb4803a6849443c1a729a03dbab4eb5aebacbbb30aa7ba5e9722f2 +size 1880117952 diff --git a/Aura-4B.i1-IQ4_XS.gguf b/Aura-4B.i1-IQ4_XS.gguf new file mode 100644 index 0000000..11600ff --- /dev/null +++ b/Aura-4B.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60bed464b23b775eb4345de3d89df9ee761585f3bcaf273463f83e4346be3302 +size 2535547584 diff --git a/Aura-4B.i1-Q2_K.gguf b/Aura-4B.i1-Q2_K.gguf new file mode 100644 index 0000000..2607e10 --- /dev/null +++ b/Aura-4B.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fede49488b61b0cb8112322b7c545292184bc429993b46850e3c749b385c7b73 +size 1839739584 diff --git a/Aura-4B.i1-Q2_K_S.gguf b/Aura-4B.i1-Q2_K_S.gguf new file mode 100644 index 0000000..5250aeb --- /dev/null +++ b/Aura-4B.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a845004404dab1713384579bd88756adc5b91151f2ee403802e4f6168c58e77d +size 1733374656 diff --git a/Aura-4B.i1-Q3_K_L.gguf b/Aura-4B.i1-Q3_K_L.gguf new file mode 100644 index 0000000..3db2ef0 --- /dev/null +++ b/Aura-4B.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e5ee68fe5e0987bc5eabd8ce4715006e738bed87de9fde53b3ff46ab80f6ea31 +size 2464860864 diff --git a/Aura-4B.i1-Q3_K_M.gguf b/Aura-4B.i1-Q3_K_M.gguf new file mode 100644 index 0000000..5e9d0a6 --- /dev/null +++ b/Aura-4B.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d24ea1df3f8b523598ba331f321b4c9e567b9300b7cfec1ddf1ff3862d3232b +size 2296564416 diff --git a/Aura-4B.i1-Q3_K_S.gguf b/Aura-4B.i1-Q3_K_S.gguf new file mode 100644 index 0000000..44c073f --- /dev/null +++ b/Aura-4B.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:110813c9e7e58770ddb10b2ace9d21e4130aa9f430606ada0bddbd08b9a84017 +size 2101529280 diff --git a/Aura-4B.i1-Q4_0.gguf b/Aura-4B.i1-Q4_0.gguf new file mode 100644 index 0000000..7b42227 --- /dev/null +++ b/Aura-4B.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4156a5eebce3ea569b5d6ba355615f63ff7d1dda9e5d7cf2855e9a36b69b2525 +size 2655601344 diff --git a/Aura-4B.i1-Q4_K_M.gguf b/Aura-4B.i1-Q4_K_M.gguf new file mode 100644 index 0000000..cba3cbd --- /dev/null +++ b/Aura-4B.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d86bb2e3b444ea7afb6c796b79d77ae2d1b23d8adc2df60be4475b9abeaabe74 +size 2778284736 diff --git a/Aura-4B.i1-Q4_K_S.gguf b/Aura-4B.i1-Q4_K_S.gguf new file mode 100644 index 0000000..8342b45 --- /dev/null +++ b/Aura-4B.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da0c9378103d96263033a38ae1d7f0d7e036e438ce94ea4efefbc76b0649ca53 +size 2664252096 diff --git a/Aura-4B.i1-Q5_K_M.gguf b/Aura-4B.i1-Q5_K_M.gguf new file mode 100644 index 0000000..23d7b2a --- /dev/null +++ b/Aura-4B.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5083cd6d6cdedc7298d00083b9932f654fdcac49d2650a56ddf98b756489543c +size 3230188224 diff --git a/Aura-4B.i1-Q5_K_S.gguf b/Aura-4B.i1-Q5_K_S.gguf new file mode 100644 index 0000000..60e3236 --- /dev/null +++ b/Aura-4B.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac033e0cf48cfe38015737e40ded6b179c1d1b65b58ddfc6e31bd596412d9da8 +size 3163341504 diff --git a/Aura-4B.i1-Q6_K.gguf b/Aura-4B.i1-Q6_K.gguf new file mode 100644 index 0000000..cb98157 --- /dev/null +++ b/Aura-4B.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8eac1d2393c16cca5f9f85d5084c5d0ec9bc692f20dece69b4a61ffb54a8f5bf +size 3710335680 diff --git a/README.md b/README.md new file mode 100644 index 0000000..a8e4993 --- /dev/null +++ b/README.md @@ -0,0 +1,79 @@ +--- +base_model: AuraIndustries/Aura-4B +datasets: +- Mielikki/Erebus-87k +- FourOhFour/Instruct_Phase +- FourOhFour/RP_Phase +- anthracite-core/full-opus-chosen-hermes-rejected-kto-v1 +language: +- en +library_name: transformers +license: apache-2.0 +quantized_by: mradermacher +--- +## About + + + + + + +weighted/imatrix quants of https://huggingface.co/AuraIndustries/Aura-4B + + +static quants are available at https://huggingface.co/mradermacher/Aura-4B-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-IQ1_S.gguf) | i1-IQ1_S | 1.3 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-IQ1_M.gguf) | i1-IQ1_M | 1.4 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 1.5 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-IQ2_XS.gguf) | i1-IQ2_XS | 1.6 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-IQ2_S.gguf) | i1-IQ2_S | 1.7 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-IQ2_M.gguf) | i1-IQ2_M | 1.8 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-Q2_K_S.gguf) | i1-Q2_K_S | 1.8 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-Q2_K.gguf) | i1-Q2_K | 1.9 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 2.0 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-IQ3_XS.gguf) | i1-IQ3_XS | 2.1 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-Q3_K_S.gguf) | i1-Q3_K_S | 2.2 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-IQ3_S.gguf) | i1-IQ3_S | 2.2 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-IQ3_M.gguf) | i1-IQ3_M | 2.3 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-Q3_K_M.gguf) | i1-Q3_K_M | 2.4 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-Q3_K_L.gguf) | i1-Q3_K_L | 2.6 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-IQ4_XS.gguf) | i1-IQ4_XS | 2.6 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-Q4_0.gguf) | i1-Q4_0 | 2.8 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-Q4_K_S.gguf) | i1-Q4_K_S | 2.8 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-Q4_K_M.gguf) | i1-Q4_K_M | 2.9 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-Q5_K_S.gguf) | i1-Q5_K_S | 3.3 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-Q5_K_M.gguf) | i1-Q5_K_M | 3.3 | | +| [GGUF](https://huggingface.co/mradermacher/Aura-4B-i1-GGUF/resolve/main/Aura-4B.i1-Q6_K.gguf) | i1-Q6_K | 3.8 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/imatrix.dat b/imatrix.dat new file mode 100644 index 0000000..5cc1102 --- /dev/null +++ b/imatrix.dat @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fa25331a27d92dedf5ce14ffe4efa3221098186ccfdf56705e0c983fda62279 +size 3677437