commit 0cb1b4f52c13329163c56e352d12b1a843b171d7 Author: ModelHub XC Date: Thu Apr 30 15:07:13 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..5bb8b91 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +imatrix.dat filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..ff382e6 --- /dev/null +++ b/README.md @@ -0,0 +1,82 @@ +--- +base_model: EpistemeAI/ReasoningCore-1B-r1-0 +language: +- en +library_name: transformers +license: llama3.2 +quantized_by: mradermacher +tags: +- text-generation-inference +- transformers +- unsloth +- llama +- trl +--- +## About + + + + + + +weighted/imatrix quants of https://huggingface.co/EpistemeAI/ReasoningCore-1B-r1-0 + + +static quants are available at https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ1_S.gguf) | i1-IQ1_S | 0.5 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ1_M.gguf) | i1-IQ1_M | 0.5 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 0.5 | | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_XS.gguf) | i1-IQ2_XS | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_S.gguf) | i1-IQ2_S | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_M.gguf) | i1-IQ2_M | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q2_K_S.gguf) | i1-Q2_K_S | 0.7 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 0.7 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q2_K.gguf) | i1-Q2_K | 0.7 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_XS.gguf) | i1-IQ3_XS | 0.7 | | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_S.gguf) | i1-Q3_K_S | 0.7 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_S.gguf) | i1-IQ3_S | 0.7 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_M.gguf) | i1-IQ3_M | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_M.gguf) | i1-Q3_K_M | 0.8 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_L.gguf) | i1-Q3_K_L | 0.8 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ4_XS.gguf) | i1-IQ4_XS | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ4_NL.gguf) | i1-IQ4_NL | 0.9 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_0.gguf) | i1-Q4_0 | 0.9 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_K_S.gguf) | i1-Q4_K_S | 0.9 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_K_M.gguf) | i1-Q4_K_M | 0.9 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_1.gguf) | i1-Q4_1 | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q5_K_S.gguf) | i1-Q5_K_S | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q5_K_M.gguf) | i1-Q5_K_M | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/ReasoningCore-Llama-3.2-1B-RE01-0-i1-GGUF/resolve/main/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q6_K.gguf) | i1-Q6_K | 1.1 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ1_M.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ1_M.gguf new file mode 100644 index 0000000..18e8763 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d5355ca4db20d590093937494b4bf15b1f3f13d03ae7f50b2bffd2f36727cb9b +size 413606912 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ1_S.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ1_S.gguf new file mode 100644 index 0000000..8cae0ea --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7dce81e55c6b358564d0c1d90fe202bcbed0847d383d52a2c18284687435adc8 +size 393552896 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_M.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_M.gguf new file mode 100644 index 0000000..46d3672 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d5095d612b5e6981f344f3564a4c03af91950a94467e054001b6f2ef156d9c7 +size 515449856 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_S.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_S.gguf new file mode 100644 index 0000000..1ead05c --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e27de4afd6b87a93b3164476a48c512828215e36d8fcd648cd5d8004453e7e8 +size 488711168 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_XS.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_XS.gguf new file mode 100644 index 0000000..cc91f9f --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c8c56c557ebe782f162b1601a90e0e3a45911fd96b3e4cf942a29f4105e90ce +size 475866112 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_XXS.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..576f561 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca63951f2b364caa6ff5919f02ff63df24ef59ecb2fc1ce90a8c1481f6932f61 +size 447030272 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_M.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_M.gguf new file mode 100644 index 0000000..a6f6e22 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:621340114a0da85aaa91bbb5b434e43d248bb84c625cd607cb863338474d34b1 +size 657290240 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_S.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_S.gguf new file mode 100644 index 0000000..c810519 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ede63f60d61d0db698119ae7acfa1170c35bde3c553ae95013779f249df4b760 +size 643920896 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_XS.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_XS.gguf new file mode 100644 index 0000000..99b3314 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fdc24972dd4a54fe28d6863fbc465adf3cd8842fde59c8f3f8f49f92e0d78e1 +size 621114368 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_XXS.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..59fe211 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65d0b77c17b6f80057cb8b271487ae4de5149eb26da26fd5e9a5d93b9afb3903 +size 562111488 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ4_NL.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ4_NL.gguf new file mode 100644 index 0000000..b1b0e9e --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e525257bc00938cc4eea87065d6e753b93f5bcf49ba7cdef1d794b00f656ffff +size 773026816 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ4_XS.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ4_XS.gguf new file mode 100644 index 0000000..6487a95 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b87ec4b0d72bc6217aafdc4018311226516bb7a26e977cfb0911e0572c9e965 +size 743142400 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q2_K.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q2_K.gguf new file mode 100644 index 0000000..9ef17b1 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4df779975ee056caf8fdd17c51fbf1529700fc80841b2b62cba03b51e24eca91 +size 580875264 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q2_K_S.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q2_K_S.gguf new file mode 100644 index 0000000..18f7c5f --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c267634aefb710a431f38e3729a438da4685b0881d772357ae9f582a7549f580 +size 554660864 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_L.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_L.gguf new file mode 100644 index 0000000..b1f940a --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57138a5835628882ac818a38b7af1e23e25bb4c32ed46c2956041292434a956e +size 732525568 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_M.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_M.gguf new file mode 100644 index 0000000..3965835 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a631edbadba501f15573f835378e2ea2b2bf68421519658d5ccccc03422e46b +size 690844672 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_S.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_S.gguf new file mode 100644 index 0000000..81f058d --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9d302d857ca1b7bf10018d67cb3580a27c55d00c33b23ebe4f427983b6a1397 +size 641692672 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_0.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_0.gguf new file mode 100644 index 0000000..a2f6bfc --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4fc7bca0396f9c61b00a99edaa8840bcb092ee2ec187e407f2b85ad23520bae +size 773026816 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_1.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_1.gguf new file mode 100644 index 0000000..35eb961 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d29814e6c4586005813b0de8f43dce96f05b9f29008f3dc65754617f6e16de3 +size 831747072 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_K_M.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_K_M.gguf new file mode 100644 index 0000000..70ca482 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e03fd8f8ccd074f74d61093472c1d6afaad986829b5e30b0e9fc9c6714b845e +size 807695360 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_K_S.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_K_S.gguf new file mode 100644 index 0000000..9ee4945 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9931499d3c34db74b84fcde7b6e26109fb875b1f1775105109e57e8e3cbf4267 +size 775648256 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q5_K_M.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q5_K_M.gguf new file mode 100644 index 0000000..f0d2a64 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e3e85d28e6e9756197f75b6ab229de9ceac26b7fe8c59467b6a1a9a895e7cb1 +size 911504384 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q5_K_S.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q5_K_S.gguf new file mode 100644 index 0000000..78ef5dc --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e1e7bb6f8b934cda9d96e3d6425631f52182d97a5e7f632e7c965a7653e8f63 +size 892564480 diff --git a/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q6_K.gguf b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q6_K.gguf new file mode 100644 index 0000000..ac81a11 --- /dev/null +++ b/ReasoningCore-Llama-3.2-1B-RE01-0.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86792dc1259cf32fb522be6f1283df3ad7a2aa08e007eae98bb76f93ee4556c8 +size 1021801472 diff --git a/imatrix.dat b/imatrix.dat new file mode 100644 index 0000000..12056ee --- /dev/null +++ b/imatrix.dat @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35f90d63697671c393e42742769efa302762788a7924563d3cd50988f054f9f0 +size 1314413