From 79056479f1943e81173ee9318437f61517046d5b Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Thu, 30 Apr 2026 15:34:56 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/SJT-2B-i1-GGUF Source: Original Platform --- .gitattributes | 60 +++++++++++++++++++++++++++++++ README.md | 82 ++++++++++++++++++++++++++++++++++++++++++ SJT-2B.i1-IQ1_M.gguf | 3 ++ SJT-2B.i1-IQ1_S.gguf | 3 ++ SJT-2B.i1-IQ2_M.gguf | 3 ++ SJT-2B.i1-IQ2_S.gguf | 3 ++ SJT-2B.i1-IQ2_XS.gguf | 3 ++ SJT-2B.i1-IQ2_XXS.gguf | 3 ++ SJT-2B.i1-IQ3_M.gguf | 3 ++ SJT-2B.i1-IQ3_S.gguf | 3 ++ SJT-2B.i1-IQ3_XS.gguf | 3 ++ SJT-2B.i1-IQ3_XXS.gguf | 3 ++ SJT-2B.i1-IQ4_NL.gguf | 3 ++ SJT-2B.i1-IQ4_XS.gguf | 3 ++ SJT-2B.i1-Q2_K.gguf | 3 ++ SJT-2B.i1-Q2_K_S.gguf | 3 ++ SJT-2B.i1-Q3_K_L.gguf | 3 ++ SJT-2B.i1-Q3_K_M.gguf | 3 ++ SJT-2B.i1-Q3_K_S.gguf | 3 ++ SJT-2B.i1-Q4_0.gguf | 3 ++ SJT-2B.i1-Q4_1.gguf | 3 ++ SJT-2B.i1-Q4_K_M.gguf | 3 ++ SJT-2B.i1-Q4_K_S.gguf | 3 ++ SJT-2B.i1-Q5_K_M.gguf | 3 ++ SJT-2B.i1-Q5_K_S.gguf | 3 ++ SJT-2B.i1-Q6_K.gguf | 3 ++ imatrix.dat | 3 ++ 27 files changed, 217 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 SJT-2B.i1-IQ1_M.gguf create mode 100644 SJT-2B.i1-IQ1_S.gguf create mode 100644 SJT-2B.i1-IQ2_M.gguf create mode 100644 SJT-2B.i1-IQ2_S.gguf create mode 100644 SJT-2B.i1-IQ2_XS.gguf create mode 100644 SJT-2B.i1-IQ2_XXS.gguf create mode 100644 SJT-2B.i1-IQ3_M.gguf create mode 100644 SJT-2B.i1-IQ3_S.gguf create mode 100644 SJT-2B.i1-IQ3_XS.gguf create mode 100644 SJT-2B.i1-IQ3_XXS.gguf create mode 100644 SJT-2B.i1-IQ4_NL.gguf create mode 100644 SJT-2B.i1-IQ4_XS.gguf create mode 100644 SJT-2B.i1-Q2_K.gguf create mode 100644 SJT-2B.i1-Q2_K_S.gguf create mode 100644 SJT-2B.i1-Q3_K_L.gguf create mode 100644 SJT-2B.i1-Q3_K_M.gguf create mode 100644 SJT-2B.i1-Q3_K_S.gguf create mode 100644 SJT-2B.i1-Q4_0.gguf create mode 100644 SJT-2B.i1-Q4_1.gguf create mode 100644 SJT-2B.i1-Q4_K_M.gguf create mode 100644 SJT-2B.i1-Q4_K_S.gguf create mode 100644 SJT-2B.i1-Q5_K_M.gguf create mode 100644 SJT-2B.i1-Q5_K_S.gguf create mode 100644 SJT-2B.i1-Q6_K.gguf create mode 100644 imatrix.dat diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..154e99d --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +imatrix.dat filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +SJT-2B.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..c5b353d --- /dev/null +++ b/README.md @@ -0,0 +1,82 @@ +--- +base_model: Sakalti/SJT-2B +language: +- en +- ja +library_name: transformers +license: gemma +quantized_by: mradermacher +tags: +- unsloth +- trl +- sft +- gemma +--- +## About + + + + + + +weighted/imatrix quants of https://huggingface.co/Sakalti/SJT-2B + + +static quants are available at https://huggingface.co/mradermacher/SJT-2B-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ1_S.gguf) | i1-IQ1_S | 0.9 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ1_M.gguf) | i1-IQ1_M | 1.0 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ2_XS.gguf) | i1-IQ2_XS | 1.1 | | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ2_S.gguf) | i1-IQ2_S | 1.1 | | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ2_M.gguf) | i1-IQ2_M | 1.2 | | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q2_K_S.gguf) | i1-Q2_K_S | 1.3 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 1.3 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q2_K.gguf) | i1-Q2_K | 1.3 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ3_XS.gguf) | i1-IQ3_XS | 1.4 | | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ3_S.gguf) | i1-IQ3_S | 1.5 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q3_K_S.gguf) | i1-Q3_K_S | 1.5 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ3_M.gguf) | i1-IQ3_M | 1.5 | | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q3_K_M.gguf) | i1-Q3_K_M | 1.6 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q3_K_L.gguf) | i1-Q3_K_L | 1.7 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ4_XS.gguf) | i1-IQ4_XS | 1.7 | | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-IQ4_NL.gguf) | i1-IQ4_NL | 1.7 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q4_0.gguf) | i1-Q4_0 | 1.7 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q4_K_S.gguf) | i1-Q4_K_S | 1.7 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q4_K_M.gguf) | i1-Q4_K_M | 1.8 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q4_1.gguf) | i1-Q4_1 | 1.9 | | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q5_K_S.gguf) | i1-Q5_K_S | 2.0 | | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q5_K_M.gguf) | i1-Q5_K_M | 2.0 | | +| [GGUF](https://huggingface.co/mradermacher/SJT-2B-i1-GGUF/resolve/main/SJT-2B.i1-Q6_K.gguf) | i1-Q6_K | 2.3 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/SJT-2B.i1-IQ1_M.gguf b/SJT-2B.i1-IQ1_M.gguf new file mode 100644 index 0000000..42e9711 --- /dev/null +++ b/SJT-2B.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7fb0e8b48f4270418c8203a1da56eae47afebb16618d16692a879bf289452961 +size 873798272 diff --git a/SJT-2B.i1-IQ1_S.gguf b/SJT-2B.i1-IQ1_S.gguf new file mode 100644 index 0000000..faaad00 --- /dev/null +++ b/SJT-2B.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68349baad87d86e6c3abf41ca0e83c622be2173b3615280f2a02513e2356de75 +size 832160384 diff --git a/SJT-2B.i1-IQ2_M.gguf b/SJT-2B.i1-IQ2_M.gguf new file mode 100644 index 0000000..cd1ec17 --- /dev/null +++ b/SJT-2B.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3de5d93c69c75d89a6c8b6fd97e28a81dd8c9ee6a17632308509f6098fbcfc3 +size 1088014976 diff --git a/SJT-2B.i1-IQ2_S.gguf b/SJT-2B.i1-IQ2_S.gguf new file mode 100644 index 0000000..9ac2c22 --- /dev/null +++ b/SJT-2B.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fe2a95ae09ea780a387bb58dee20166ab436a94fef77c80247b96150dcadff0 +size 1032497792 diff --git a/SJT-2B.i1-IQ2_XS.gguf b/SJT-2B.i1-IQ2_XS.gguf new file mode 100644 index 0000000..3c62bd5 --- /dev/null +++ b/SJT-2B.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55e62f6711e2510a651c9e01f4a980345dc150eb58b59cb7d072f256997249d5 +size 1002545792 diff --git a/SJT-2B.i1-IQ2_XXS.gguf b/SJT-2B.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..1bde720 --- /dev/null +++ b/SJT-2B.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c61fcae0d0a0f1f7b18dd213ec46718fc50f8393289221d8c0105107fd911e0 +size 943194752 diff --git a/SJT-2B.i1-IQ3_M.gguf b/SJT-2B.i1-IQ3_M.gguf new file mode 100644 index 0000000..7d6d145 --- /dev/null +++ b/SJT-2B.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f6c50c922884d76452c6babdc3cec70fb75c48e33a05d93474b073574805e0b +size 1393562240 diff --git a/SJT-2B.i1-IQ3_S.gguf b/SJT-2B.i1-IQ3_S.gguf new file mode 100644 index 0000000..926fead --- /dev/null +++ b/SJT-2B.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:784d74d78884729e440f540b77a3a785a8e3d4092ad8c684c70cd6efeefd472c +size 1360661120 diff --git a/SJT-2B.i1-IQ3_XS.gguf b/SJT-2B.i1-IQ3_XS.gguf new file mode 100644 index 0000000..fe83694 --- /dev/null +++ b/SJT-2B.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d134de5c190d231ee5bf793c108ec63d6b03b85345cbefa0984b6b896048bcc +size 1314212480 diff --git a/SJT-2B.i1-IQ3_XXS.gguf b/SJT-2B.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..0b8f9ab --- /dev/null +++ b/SJT-2B.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9dea9fbfb01eb30d84531e89acabb3e199284db6a64675b3dfd4d17d576fcc0 +size 1181686400 diff --git a/SJT-2B.i1-IQ4_NL.gguf b/SJT-2B.i1-IQ4_NL.gguf new file mode 100644 index 0000000..19c6ab3 --- /dev/null +++ b/SJT-2B.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62b2380f0b1a939e3f88172330b543b7d6545ba8e4e03f7775b1257de22dcc68 +size 1629510272 diff --git a/SJT-2B.i1-IQ4_XS.gguf b/SJT-2B.i1-IQ4_XS.gguf new file mode 100644 index 0000000..ca38160 --- /dev/null +++ b/SJT-2B.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe4ec6ed25abfe1df72492603d9d071b875aba81bec82a20ca2260b978306530 +size 1566251648 diff --git a/SJT-2B.i1-Q2_K.gguf b/SJT-2B.i1-Q2_K.gguf new file mode 100644 index 0000000..0fbe664 --- /dev/null +++ b/SJT-2B.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ede18be90f6853ae2454fcb1145322c55b02763e891d14b48642e95fd056f82 +size 1229830784 diff --git a/SJT-2B.i1-Q2_K_S.gguf b/SJT-2B.i1-Q2_K_S.gguf new file mode 100644 index 0000000..9d8589a --- /dev/null +++ b/SJT-2B.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e506375c70138da701bae211707ec254d90312e6fb66729e025ad8f7303fcdd5 +size 1170000512 diff --git a/SJT-2B.i1-Q3_K_L.gguf b/SJT-2B.i1-Q3_K_L.gguf new file mode 100644 index 0000000..15627fb --- /dev/null +++ b/SJT-2B.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:746b1b5dd3615433e365a3fd266a0b6db106c7631fbcf687b5235546e4318ecc +size 1550436992 diff --git a/SJT-2B.i1-Q3_K_M.gguf b/SJT-2B.i1-Q3_K_M.gguf new file mode 100644 index 0000000..a572508 --- /dev/null +++ b/SJT-2B.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bf63feac523ecd3c8421db7296080091e8cf35eabe0c05fd1dd195fd6d2600e +size 1461668480 diff --git a/SJT-2B.i1-Q3_K_S.gguf b/SJT-2B.i1-Q3_K_S.gguf new file mode 100644 index 0000000..2f7e021 --- /dev/null +++ b/SJT-2B.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c59219238a09cdef687cb98ddbfa279d51db5251151e343cd5da2c07d8f6aa3 +size 1360661120 diff --git a/SJT-2B.i1-Q4_0.gguf b/SJT-2B.i1-Q4_0.gguf new file mode 100644 index 0000000..8d65cf6 --- /dev/null +++ b/SJT-2B.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8726c75b77205bc5410e3e4adb60920b917cc3b22e0ca66ca78ab8e767c910f9 +size 1633491584 diff --git a/SJT-2B.i1-Q4_1.gguf b/SJT-2B.i1-Q4_1.gguf new file mode 100644 index 0000000..33aa850 --- /dev/null +++ b/SJT-2B.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28ef8824bf16fac97b4a0988828d3dc13a9f36243e21bc8f4c7df6d059dbaeb2 +size 1756027520 diff --git a/SJT-2B.i1-Q4_K_M.gguf b/SJT-2B.i1-Q4_K_M.gguf new file mode 100644 index 0000000..324ca70 --- /dev/null +++ b/SJT-2B.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae107e314d48e608ff8c71ab2bda5aa718311a389625d27149d8bf9d4d85aab7 +size 1708583552 diff --git a/SJT-2B.i1-Q4_K_S.gguf b/SJT-2B.i1-Q4_K_S.gguf new file mode 100644 index 0000000..5de6336 --- /dev/null +++ b/SJT-2B.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a29ea61345c63d9e7b4844ab8f807312227d4c3ff7b0c518462c5f9bc77c2fff +size 1638652544 diff --git a/SJT-2B.i1-Q5_K_M.gguf b/SJT-2B.i1-Q5_K_M.gguf new file mode 100644 index 0000000..601a389 --- /dev/null +++ b/SJT-2B.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1eac4a02802fd53c10565293ed8082b167e73de5ea95b094bd89daed9ad2c2b2 +size 1923279488 diff --git a/SJT-2B.i1-Q5_K_S.gguf b/SJT-2B.i1-Q5_K_S.gguf new file mode 100644 index 0000000..c75303f --- /dev/null +++ b/SJT-2B.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce51b9f237604bb92cf16a3a94e67bd9c9e1b1c744b5cee582e1dcf6c86a792d +size 1882544768 diff --git a/SJT-2B.i1-Q6_K.gguf b/SJT-2B.i1-Q6_K.gguf new file mode 100644 index 0000000..aa4233b --- /dev/null +++ b/SJT-2B.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5dc00b3530838ce7416ee514d04d48ad525d91cb697b1a36058235536384599 +size 2151393920 diff --git a/imatrix.dat b/imatrix.dat new file mode 100644 index 0000000..6b84471 --- /dev/null +++ b/imatrix.dat @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06adf641a88bf9a567facbbffcfead004d050a62326d7000f332162b3a50a6c9 +size 2375559