commit b5f3707d83df22376407ec3446b8b01ca6df18db Author: ModelHub XC Date: Sat Apr 11 05:40:55 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: gss1147/Llama-3.2-OctoThinker-iNano-1B-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..439e577 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,38 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Llama-3.2-OctoThinker-iNano-1B.f16.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Llama-3.2-OctoThinker-iNano-1B.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf b/Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf new file mode 100644 index 0000000..2153e76 --- /dev/null +++ b/Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3bededa17495fa463b6d6c98d18dd4e1bc99d615a80000693d48e13d685ce654 +size 955442560 diff --git a/Llama-3.2-OctoThinker-iNano-1B.Q5_K_M.gguf b/Llama-3.2-OctoThinker-iNano-1B.Q5_K_M.gguf new file mode 100644 index 0000000..84f9c99 --- /dev/null +++ b/Llama-3.2-OctoThinker-iNano-1B.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d17ed5f4ae3132cf20c9f32c51188a43d5dd71a0561ee1321dab2331f0f6f61 +size 1092085120 diff --git a/Llama-3.2-OctoThinker-iNano-1B.f16.gguf b/Llama-3.2-OctoThinker-iNano-1B.f16.gguf new file mode 100644 index 0000000..bca4c41 --- /dev/null +++ b/Llama-3.2-OctoThinker-iNano-1B.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e70b169a0d19e851bfe202055ce6e7079530d5451fad7df77fa276da3517618e +size 3004929408 diff --git a/README.md b/README.md new file mode 100644 index 0000000..234746e --- /dev/null +++ b/README.md @@ -0,0 +1,76 @@ +--- +license: other +library_name: gguf +tags: + - gguf + - llama + - llama-3.2 + - text-generation + - local-llm + - llama-cpp + - lm-studio + - ollama + - 1b +pipeline_tag: text-generation +language: + - en +base_model_relation: quantized +base_model: + - gss1147/Llama-3.2-OctoThinker-iNano-1B +--- + +# Llama-3.2-OctoThinker-iNano-1B-GGUF + +## Model Summary + +**Llama-3.2-OctoThinker-iNano-1B-GGUF** is the GGUF quantized release of the main model: + +**Main model repo:** +https://huggingface.co/gss1147/Llama-3.2-OctoThinker-iNano-1B + +This repository packages the model for efficient **local inference** in GGUF-compatible runtimes such as **llama.cpp**, **LM Studio**, and similar local tools. + +## GGUF to Main Model Link + +This GGUF repository corresponds to the main model repo: + +**[`gss1147/Llama-3.2-OctoThinker-iNano-1B`](https://huggingface.co/gss1147/Llama-3.2-OctoThinker-iNano-1B)** + +If you want the original non-GGUF model, training/merge details, tokenizer files, and main repository metadata, use the repo above. + +## Available Files + +This GGUF repository currently includes: + +- **Q4_K_M** — 955 MB +- **Q5_K_M** — 1.09 GB +- **F16** — 3 GB + +## Architecture + +- **Architecture:** llama +- **Model size:** 1B params + +## Intended Use + +This model is intended for: + +- local text generation +- assistant-style prompting +- lightweight reasoning tasks +- summarization +- simple coding help +- offline/local inference workflows + +## Quantization Notes + +Choose the file that best matches your hardware: + +- **Q4_K_M** for smaller size and lighter RAM usage +- **Q5_K_M** for a stronger quality-to-size balance +- **F16** for the highest-fidelity file in this repo, with much higher memory requirements + +## Example llama.cpp Usage + +```bash +llama-cli -m /path/to/Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf -p "Explain recursion in Python with a simple example." \ No newline at end of file