From 40d37d1afeb69f0d0ad99fcb4a3edbe1c1b44f0a Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Mon, 4 May 2026 12:05:45 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF Source: Original Platform --- .gitattributes | 60 ++++++++++++++++ Llama-3.1-8B-Fusion-5050.i1-IQ1_M.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-IQ1_S.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-IQ2_M.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-IQ2_S.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-IQ2_XS.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-IQ2_XXS.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-IQ3_M.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-IQ3_S.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-IQ3_XS.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-IQ3_XXS.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-IQ4_NL.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-IQ4_XS.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q2_K.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q2_K_S.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q3_K_L.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q3_K_M.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q3_K_S.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q4_0.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q4_1.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q4_K_M.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q4_K_S.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q5_K_M.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q5_K_S.gguf | 3 + Llama-3.1-8B-Fusion-5050.i1-Q6_K.gguf | 3 + README.md | 88 ++++++++++++++++++++++++ imatrix.dat | 3 + 27 files changed, 223 insertions(+) create mode 100644 .gitattributes create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ1_M.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ1_S.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ2_M.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ2_S.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ2_XS.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ2_XXS.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ3_M.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ3_S.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ3_XS.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ3_XXS.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ4_NL.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-IQ4_XS.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q2_K.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q2_K_S.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q3_K_L.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q3_K_M.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q3_K_S.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q4_0.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q4_1.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q4_K_M.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q4_K_S.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q5_K_M.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q5_K_S.gguf create mode 100644 Llama-3.1-8B-Fusion-5050.i1-Q6_K.gguf create mode 100644 README.md create mode 100644 imatrix.dat diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..daac9ce --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +imatrix.dat filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.1-8B-Fusion-5050.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ1_M.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ1_M.gguf new file mode 100644 index 0000000..8b5c8f7 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f252b66e9759d670abd5cf1202af2f7b9d3ddebbb1732ca17a57e9d949c47024 +size 2161973216 diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ1_S.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ1_S.gguf new file mode 100644 index 0000000..78d6872 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d1a59d798153365aac371a504ea578b8699a89149e76aa627d16f0280bccf05 +size 2019629024 diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ2_M.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ2_M.gguf new file mode 100644 index 0000000..15e0116 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91029b8cb1bbdafdfc854da96de357d719bd15559c2b404efcc54c1c160293fa +size 2948282336 diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ2_S.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ2_S.gguf new file mode 100644 index 0000000..e672dbe --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4dea8cc79a970b6fbdaf3514607a9b2ac4978e71332800d98dd275c8f49564a4 +size 2758490080 diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ2_XS.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ2_XS.gguf new file mode 100644 index 0000000..5be831c --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:af2b5f84c6382a6cf16e0943a9bb595db6bc1d2a8525fdbc2a830c3f4e3f2a18 +size 2605783008 diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ2_XXS.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..28f64d7 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93bb53135c5b205a24726667eb1377eebdb67fde583024dca39440d1d754b4d9 +size 2399213536 diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ3_M.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ3_M.gguf new file mode 100644 index 0000000..ef1881f --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab3de9aa93d4076c4dff4fb136edb64da0e1fb6964ff28989d79d92579a20059 +size 3784824800 diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ3_S.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ3_S.gguf new file mode 100644 index 0000000..c89927c --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27fe2153d460b2ee020ff463afbbe607b8b60b55d8a4b9fc56120e789d462954 +size 3682326496 diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ3_XS.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ3_XS.gguf new file mode 100644 index 0000000..edb7289 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef221f3e42eb78f937df9c5c4463245080eb014d73bd30b14a24723138081bd6 +size 3518748640 diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ3_XXS.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..ac1555d --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e87447a35f92b53004845e5e342a324c6cd061f7b02c25d0359a9bb686c05d5 +size 3274913760 diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ4_NL.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ4_NL.gguf new file mode 100644 index 0000000..8011ab7 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:197cc721d2492c75401748c5eccb370ed5e1fed28fc7520ce5fa001762ae754c +size 4677990368 diff --git a/Llama-3.1-8B-Fusion-5050.i1-IQ4_XS.gguf b/Llama-3.1-8B-Fusion-5050.i1-IQ4_XS.gguf new file mode 100644 index 0000000..584b7af --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45ad5580e5c58a7d03b9c61ba87b365f4274332c579b1a9c698f94d5f67f2f36 +size 4447664096 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q2_K.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q2_K.gguf new file mode 100644 index 0000000..c200346 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:268837c295f4487c638861243ca0a71e0e1a1e7052bf22d7c7eebaa666bfb20f +size 3179132896 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q2_K_S.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q2_K_S.gguf new file mode 100644 index 0000000..6301f35 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ca3959fce949e030ca2962eac3ac2e65d458a27202778fcd5e3b167f95efb92 +size 2988816352 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q3_K_L.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q3_K_L.gguf new file mode 100644 index 0000000..72bdb72 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1335ce1008efef86bcb0a559e367fc72c9e5b3e3086932fc3bdfab8b82e6077e +size 4321957856 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q3_K_M.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q3_K_M.gguf new file mode 100644 index 0000000..a9bcc34 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:223dff0c0cb03285c1797ee7cea451b3ffd6d4ec13762e3d7d988a2bef5daf9f +size 4018919392 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q3_K_S.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q3_K_S.gguf new file mode 100644 index 0000000..34064d0 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da109c23b408fa400aa5448aa9af27d0d658d8839fb3570bba8c0610b7ca325c +size 3664500704 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q4_0.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q4_0.gguf new file mode 100644 index 0000000..49e4b60 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:894a07f6f4c97963321bd64925b1e2b65a06768a559b3e30ed475c8c20a47df5 +size 4675893216 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q4_1.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q4_1.gguf new file mode 100644 index 0000000..ccd8092 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:842e2bde24a0bc54b81b8002a72344745c5cc7e85609f63def4e94cdbf156975 +size 5130254304 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q4_K_M.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q4_K_M.gguf new file mode 100644 index 0000000..94de726 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be9b2050d2c9805464a99c51aaa23a14ba1c9cf105274df10340ae1a63fb0aba +size 4920735712 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q4_K_S.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q4_K_S.gguf new file mode 100644 index 0000000..dc9b157 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fe0062790237fa6b1c8d717a6f82c591cb9365cafcbf62dff946dd6f6f173cb +size 4692670432 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q5_K_M.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q5_K_M.gguf new file mode 100644 index 0000000..340bbc2 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4cff9144dc69393247853e406168bc6ff1651944898edfa2725da9d11b6ad972 +size 5732988896 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q5_K_S.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q5_K_S.gguf new file mode 100644 index 0000000..49b0c7d --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c33e4bb1e75f3024a9543ff0d056a890ddcf7640bdf8860f3783956874b1e27 +size 5599295456 diff --git a/Llama-3.1-8B-Fusion-5050.i1-Q6_K.gguf b/Llama-3.1-8B-Fusion-5050.i1-Q6_K.gguf new file mode 100644 index 0000000..3d2dc31 --- /dev/null +++ b/Llama-3.1-8B-Fusion-5050.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a179b3b85839a24b3f6d63c8f7cc71c7f5cbec1e84cedea9c92f39b2daac30b7 +size 6596007904 diff --git a/README.md b/README.md new file mode 100644 index 0000000..675203d --- /dev/null +++ b/README.md @@ -0,0 +1,88 @@ +--- +base_model: huihui-ai/Llama-3.1-8B-Fusion-5050 +language: +- en +library_name: transformers +license: llama3.1 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- Text Generation +- llama3.1 +- text-generation-inference +- Inference Endpoints +- Transformers +- Fusion +--- +## About + + + + + + +weighted/imatrix quants of https://huggingface.co/huihui-ai/Llama-3.1-8B-Fusion-5050 + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Llama-3.1-8B-Fusion-5050-i1-GGUF).*** + +static quants are available at https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ1_S.gguf) | i1-IQ1_S | 2.1 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ1_M.gguf) | i1-IQ1_M | 2.3 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 2.5 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ2_XS.gguf) | i1-IQ2_XS | 2.7 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ2_S.gguf) | i1-IQ2_S | 2.9 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ2_M.gguf) | i1-IQ2_M | 3.0 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q2_K_S.gguf) | i1-Q2_K_S | 3.1 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q2_K.gguf) | i1-Q2_K | 3.3 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 3.4 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ3_XS.gguf) | i1-IQ3_XS | 3.6 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q3_K_S.gguf) | i1-Q3_K_S | 3.8 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ3_S.gguf) | i1-IQ3_S | 3.8 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ3_M.gguf) | i1-IQ3_M | 3.9 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q3_K_M.gguf) | i1-Q3_K_M | 4.1 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q3_K_L.gguf) | i1-Q3_K_L | 4.4 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ4_XS.gguf) | i1-IQ4_XS | 4.5 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q4_0.gguf) | i1-Q4_0 | 4.8 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-IQ4_NL.gguf) | i1-IQ4_NL | 4.8 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q4_K_S.gguf) | i1-Q4_K_S | 4.8 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q4_K_M.gguf) | i1-Q4_K_M | 5.0 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q4_1.gguf) | i1-Q4_1 | 5.2 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q5_K_S.gguf) | i1-Q5_K_S | 5.7 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q5_K_M.gguf) | i1-Q5_K_M | 5.8 | | +| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Fusion-5050-i1-GGUF/resolve/main/Llama-3.1-8B-Fusion-5050.i1-Q6_K.gguf) | i1-Q6_K | 6.7 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/imatrix.dat b/imatrix.dat new file mode 100644 index 0000000..b763159 --- /dev/null +++ b/imatrix.dat @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:906de9ce8a05ca14e7f49fd2d745a1ab51a7695b6d28bc2f8c7be7f80ddf8058 +size 4988157