From 2fc1cbcbad3f66cbbc9360126bb92a4c5a7bf8cb Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Fri, 10 Apr 2026 19:18:57 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF Source: Original Platform --- .gitattributes | 60 ++++++++++++++++ Oganesson-TinyLlama-1.2B.i1-IQ1_M.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-IQ1_S.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-IQ2_M.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-IQ2_S.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-IQ2_XS.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-IQ2_XXS.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-IQ3_M.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-IQ3_S.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-IQ3_XS.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-IQ3_XXS.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-IQ4_NL.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-IQ4_XS.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q2_K.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q2_K_S.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q3_K_L.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q3_K_M.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q3_K_S.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q4_0.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q4_1.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q4_K_M.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q4_K_S.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q5_K_M.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q5_K_S.gguf | 3 + Oganesson-TinyLlama-1.2B.i1-Q6_K.gguf | 3 + Oganesson-TinyLlama-1.2B.imatrix.gguf | 3 + README.md | 91 ++++++++++++++++++++++++ 27 files changed, 226 insertions(+) create mode 100644 .gitattributes create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ1_M.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ1_S.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ2_M.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ2_S.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ2_XS.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ2_XXS.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ3_M.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ3_S.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ3_XS.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ3_XXS.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ4_NL.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-IQ4_XS.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q2_K.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q2_K_S.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q3_K_L.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q3_K_M.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q3_K_S.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q4_0.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q4_1.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q4_K_M.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q4_K_S.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q5_K_M.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q5_K_S.gguf create mode 100644 Oganesson-TinyLlama-1.2B.i1-Q6_K.gguf create mode 100644 Oganesson-TinyLlama-1.2B.imatrix.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..6249955 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.imatrix.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Oganesson-TinyLlama-1.2B.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ1_M.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ1_M.gguf new file mode 100644 index 0000000..b3caca3 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d7fd6c8ae025c39e8d1a63305b3a7a238d6cc3a43868fa3ea0bbfe00674b92d +size 413603456 diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ1_S.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ1_S.gguf new file mode 100644 index 0000000..f63a4bc --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:954d82e263ef324345e56182bb1e9c0bd98bdba0b0358f107084017f1ca407b9 +size 393549440 diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ2_M.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ2_M.gguf new file mode 100644 index 0000000..df90cb5 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a89ca1a23f13928fedef4418f6d30dc8fbe280698d0593feb11a5a41ba3715e1 +size 515446400 diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ2_S.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ2_S.gguf new file mode 100644 index 0000000..2e4acdf --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7de90574c05b5be43ab3e85d1ab739bfbb04f216eec433f03df49739fc55b667 +size 488707712 diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ2_XS.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ2_XS.gguf new file mode 100644 index 0000000..a40da9c --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d23a43a55d76b72bd4a6f04626cf43960f16f7337ea205a0070c4b07b60c3fbe +size 475862656 diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ2_XXS.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..8418285 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ecff442f69c72ccc2ae665fab1471801d0f1c8ab22832f250f6f050dcdf4d13 +size 447026816 diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ3_M.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ3_M.gguf new file mode 100644 index 0000000..58a38cc --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c37ffe5ba378bf4010eee318d78915b1f388e64a35fb7bae654ce5348c121c12 +size 657286784 diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ3_S.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ3_S.gguf new file mode 100644 index 0000000..cc7c4cd --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ef4a32495c398c436aa63a396701e63d2f7ff759269bc472c8ecfa5df908b23 +size 643917440 diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ3_XS.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ3_XS.gguf new file mode 100644 index 0000000..7026f0a --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9191cfdb2e68b27575f7a7cdeaee34d35a65ec3dbcfeae516b54c438cc82d23c +size 621110912 diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ3_XXS.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..dd0a335 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41693f45c823c8bd560c4833311c4af0fc49ee2ea4cc53f9e2fd26fee4a557e4 +size 562108032 diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ4_NL.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ4_NL.gguf new file mode 100644 index 0000000..e5165c4 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f727ff854d782e77150aacc742cf4efa6cb47e2bb2ac0c40ad3d514e8c0633fa +size 773023360 diff --git a/Oganesson-TinyLlama-1.2B.i1-IQ4_XS.gguf b/Oganesson-TinyLlama-1.2B.i1-IQ4_XS.gguf new file mode 100644 index 0000000..5d61118 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a381aa056ecd1c2e6fc620ea323f9993a2eb4685a3c9417d3bd246c318586b52 +size 743138944 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q2_K.gguf b/Oganesson-TinyLlama-1.2B.i1-Q2_K.gguf new file mode 100644 index 0000000..f6f8afa --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a94a2693693f17004a234db2ac9098acd9c3c6bbe70f48ee02cb4dbcd9f0b66e +size 580871808 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q2_K_S.gguf b/Oganesson-TinyLlama-1.2B.i1-Q2_K_S.gguf new file mode 100644 index 0000000..e856082 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:450a6c23969383c9b6e9269e12ef78071277720826d6053349c47f05a3e786f0 +size 554657408 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q3_K_L.gguf b/Oganesson-TinyLlama-1.2B.i1-Q3_K_L.gguf new file mode 100644 index 0000000..518669a --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34d2216a268b9f5674d291ab1307cb60a5c10de5bc54eaf48a217b309333cc95 +size 732522112 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q3_K_M.gguf b/Oganesson-TinyLlama-1.2B.i1-Q3_K_M.gguf new file mode 100644 index 0000000..4da9cfd --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7aca8e817aec3fc5c432bdc9c2218b657b36a3500e4a17a45437774f89a53726 +size 690841216 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q3_K_S.gguf b/Oganesson-TinyLlama-1.2B.i1-Q3_K_S.gguf new file mode 100644 index 0000000..a911ec1 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a953e340530923852d8ca75bdd66bc9af16118772b11dc925076f8ea1e767169 +size 641689216 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q4_0.gguf b/Oganesson-TinyLlama-1.2B.i1-Q4_0.gguf new file mode 100644 index 0000000..078c809 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff26ac1a7773d93caa5d953c41210ba8ccc995d47e116e3993adb60f8b80245e +size 773023360 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q4_1.gguf b/Oganesson-TinyLlama-1.2B.i1-Q4_1.gguf new file mode 100644 index 0000000..f1f4958 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7b4854c08a82b62b0d020608b3764aaf173e0d6f4986ff83e07ddbe24e1185d +size 831743616 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q4_K_M.gguf b/Oganesson-TinyLlama-1.2B.i1-Q4_K_M.gguf new file mode 100644 index 0000000..7a3b8a5 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:777668ff5bfe710f674aaf3ee01d6d140ec0cb68b58714aad8487ff958b10433 +size 807691904 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q4_K_S.gguf b/Oganesson-TinyLlama-1.2B.i1-Q4_K_S.gguf new file mode 100644 index 0000000..e1df34a --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fbe2c813f1033a5244cc3807b269fd619e2776703616cc4f63d0aee8172fcb0 +size 775644800 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q5_K_M.gguf b/Oganesson-TinyLlama-1.2B.i1-Q5_K_M.gguf new file mode 100644 index 0000000..fcdc5da --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2999a2872545ee77e61e9cf2504e8916e6355308135e6bc4af646654b8413e7 +size 911500928 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q5_K_S.gguf b/Oganesson-TinyLlama-1.2B.i1-Q5_K_S.gguf new file mode 100644 index 0000000..56d641a --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6ee92bc3e55e79381590a5371c85e714865e5b422bf37e9578c46a0c1bd9838 +size 892561024 diff --git a/Oganesson-TinyLlama-1.2B.i1-Q6_K.gguf b/Oganesson-TinyLlama-1.2B.i1-Q6_K.gguf new file mode 100644 index 0000000..e18d8b2 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:882e95bd0885d1ef45106be8e018d258ff0588f7c44fa0d24a6719e6e1c298f0 +size 1021798016 diff --git a/Oganesson-TinyLlama-1.2B.imatrix.gguf b/Oganesson-TinyLlama-1.2B.imatrix.gguf new file mode 100644 index 0000000..615f182 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.imatrix.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b6f03d26518ba1a38b157e82403c7dc17a1e9626b5b0edbb0bcedb4c48d1cee +size 1328000 diff --git a/README.md b/README.md new file mode 100644 index 0000000..6b94df3 --- /dev/null +++ b/README.md @@ -0,0 +1,91 @@ +--- +base_model: prithivMLmods/Oganesson-TinyLlama-1.2B +language: +- en +library_name: transformers +license: llama3.2 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- text-generation-inference +- code +- llama-3.2 +- math +- general-purpose +--- +## About + + + + + + + + + +weighted/imatrix quants of https://huggingface.co/prithivMLmods/Oganesson-TinyLlama-1.2B + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Oganesson-TinyLlama-1.2B-i1-GGUF).*** + +static quants are available at https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.imatrix.gguf) | imatrix | 0.1 | imatrix file (for creating your own quants) | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ1_S.gguf) | i1-IQ1_S | 0.5 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ1_M.gguf) | i1-IQ1_M | 0.5 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 0.5 | | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ2_XS.gguf) | i1-IQ2_XS | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ2_S.gguf) | i1-IQ2_S | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ2_M.gguf) | i1-IQ2_M | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q2_K_S.gguf) | i1-Q2_K_S | 0.7 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 0.7 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q2_K.gguf) | i1-Q2_K | 0.7 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ3_XS.gguf) | i1-IQ3_XS | 0.7 | | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q3_K_S.gguf) | i1-Q3_K_S | 0.7 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ3_S.gguf) | i1-IQ3_S | 0.7 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ3_M.gguf) | i1-IQ3_M | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q3_K_M.gguf) | i1-Q3_K_M | 0.8 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q3_K_L.gguf) | i1-Q3_K_L | 0.8 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ4_XS.gguf) | i1-IQ4_XS | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-IQ4_NL.gguf) | i1-IQ4_NL | 0.9 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q4_0.gguf) | i1-Q4_0 | 0.9 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q4_K_S.gguf) | i1-Q4_K_S | 0.9 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q4_K_M.gguf) | i1-Q4_K_M | 0.9 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q4_1.gguf) | i1-Q4_1 | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q5_K_S.gguf) | i1-Q5_K_S | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q5_K_M.gguf) | i1-Q5_K_M | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/Oganesson-TinyLlama-1.2B-i1-GGUF/resolve/main/Oganesson-TinyLlama-1.2B.i1-Q6_K.gguf) | i1-Q6_K | 1.1 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + +