From b5f3707d83df22376407ec3446b8b01ca6df18db Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sat, 11 Apr 2026 05:40:55 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: gss1147/Llama-3.2-OctoThinker-iNano-1B-GGUF Source: Original Platform --- .gitattributes | 38 +++++++++++ Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf | 3 + Llama-3.2-OctoThinker-iNano-1B.Q5_K_M.gguf | 3 + Llama-3.2-OctoThinker-iNano-1B.f16.gguf | 3 + README.md | 76 ++++++++++++++++++++++ 5 files changed, 123 insertions(+) create mode 100644 .gitattributes create mode 100644 Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf create mode 100644 Llama-3.2-OctoThinker-iNano-1B.Q5_K_M.gguf create mode 100644 Llama-3.2-OctoThinker-iNano-1B.f16.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..439e577 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,38 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Llama-3.2-OctoThinker-iNano-1B.f16.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-OctoThinker-iNano-1B.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf b/Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf new file mode 100644 index 0000000..2153e76 --- /dev/null +++ b/Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3bededa17495fa463b6d6c98d18dd4e1bc99d615a80000693d48e13d685ce654 +size 955442560 diff --git a/Llama-3.2-OctoThinker-iNano-1B.Q5_K_M.gguf b/Llama-3.2-OctoThinker-iNano-1B.Q5_K_M.gguf new file mode 100644 index 0000000..84f9c99 --- /dev/null +++ b/Llama-3.2-OctoThinker-iNano-1B.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d17ed5f4ae3132cf20c9f32c51188a43d5dd71a0561ee1321dab2331f0f6f61 +size 1092085120 diff --git a/Llama-3.2-OctoThinker-iNano-1B.f16.gguf b/Llama-3.2-OctoThinker-iNano-1B.f16.gguf new file mode 100644 index 0000000..bca4c41 --- /dev/null +++ b/Llama-3.2-OctoThinker-iNano-1B.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e70b169a0d19e851bfe202055ce6e7079530d5451fad7df77fa276da3517618e +size 3004929408 diff --git a/README.md b/README.md new file mode 100644 index 0000000..234746e --- /dev/null +++ b/README.md @@ -0,0 +1,76 @@ +--- +license: other +library_name: gguf +tags: + - gguf + - llama + - llama-3.2 + - text-generation + - local-llm + - llama-cpp + - lm-studio + - ollama + - 1b +pipeline_tag: text-generation +language: + - en +base_model_relation: quantized +base_model: + - gss1147/Llama-3.2-OctoThinker-iNano-1B +--- + +# Llama-3.2-OctoThinker-iNano-1B-GGUF + +## Model Summary + +**Llama-3.2-OctoThinker-iNano-1B-GGUF** is the GGUF quantized release of the main model: + +**Main model repo:** +https://huggingface.co/gss1147/Llama-3.2-OctoThinker-iNano-1B + +This repository packages the model for efficient **local inference** in GGUF-compatible runtimes such as **llama.cpp**, **LM Studio**, and similar local tools. + +## GGUF to Main Model Link + +This GGUF repository corresponds to the main model repo: + +**[`gss1147/Llama-3.2-OctoThinker-iNano-1B`](https://huggingface.co/gss1147/Llama-3.2-OctoThinker-iNano-1B)** + +If you want the original non-GGUF model, training/merge details, tokenizer files, and main repository metadata, use the repo above. + +## Available Files + +This GGUF repository currently includes: + +- **Q4_K_M** — 955 MB +- **Q5_K_M** — 1.09 GB +- **F16** — 3 GB + +## Architecture + +- **Architecture:** llama +- **Model size:** 1B params + +## Intended Use + +This model is intended for: + +- local text generation +- assistant-style prompting +- lightweight reasoning tasks +- summarization +- simple coding help +- offline/local inference workflows + +## Quantization Notes + +Choose the file that best matches your hardware: + +- **Q4_K_M** for smaller size and lighter RAM usage +- **Q5_K_M** for a stronger quality-to-size balance +- **F16** for the highest-fidelity file in this repo, with much higher memory requirements + +## Example llama.cpp Usage + +```bash +llama-cli -m /path/to/Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf -p "Explain recursion in Python with a simple example." \ No newline at end of file