From c9c3093f1a910fdd06330c89e0f55a8ad426f38c Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Wed, 22 Apr 2026 01:40:51 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: prithivMLmods/Oganesson-TinyLlama-1.2B-GGUF Source: Original Platform --- .gitattributes | 47 ++++++++++++++++++++++++++++ Oganesson-TinyLlama-1.2B.BF16.gguf | 3 ++ Oganesson-TinyLlama-1.2B.F16.gguf | 3 ++ Oganesson-TinyLlama-1.2B.F32.gguf | 3 ++ Oganesson-TinyLlama-1.2B.Q4_K_M.gguf | 3 ++ README.md | 38 ++++++++++++++++++++++ config.json | 3 ++ configuration.json | 1 + 8 files changed, 101 insertions(+) create mode 100644 .gitattributes create mode 100644 Oganesson-TinyLlama-1.2B.BF16.gguf create mode 100644 Oganesson-TinyLlama-1.2B.F16.gguf create mode 100644 Oganesson-TinyLlama-1.2B.F32.gguf create mode 100644 Oganesson-TinyLlama-1.2B.Q4_K_M.gguf create mode 100644 README.md create mode 100644 config.json create mode 100644 configuration.json diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..53d7257 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bin.* filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zstandard filter=lfs diff=lfs merge=lfs -text +*.tfevents* filter=lfs diff=lfs merge=lfs -text +*.db* filter=lfs diff=lfs merge=lfs -text +*.ark* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.gguf* filter=lfs diff=lfs merge=lfs -text +*.ggml filter=lfs diff=lfs merge=lfs -text +*.llamafile* filter=lfs diff=lfs merge=lfs -text +*.pt2 filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text \ No newline at end of file diff --git a/Oganesson-TinyLlama-1.2B.BF16.gguf b/Oganesson-TinyLlama-1.2B.BF16.gguf new file mode 100644 index 0000000..020725b --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.BF16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69f7e6de645e44a327247f82d7de459b5557e75b0652395cf8f9f1ef9131e859 +size 2479591680 diff --git a/Oganesson-TinyLlama-1.2B.F16.gguf b/Oganesson-TinyLlama-1.2B.F16.gguf new file mode 100644 index 0000000..57f7a18 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.F16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3008d0f6bf06dff1c5179e661d4722108af9862d423e358ceefc88e38cd42ab0 +size 2479591680 diff --git a/Oganesson-TinyLlama-1.2B.F32.gguf b/Oganesson-TinyLlama-1.2B.F32.gguf new file mode 100644 index 0000000..353752c --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.F32.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40af97bedd430106e36a8b783c329e4e856be9605817c72c20f55dbc25fd74e2 +size 4951085312 diff --git a/Oganesson-TinyLlama-1.2B.Q4_K_M.gguf b/Oganesson-TinyLlama-1.2B.Q4_K_M.gguf new file mode 100644 index 0000000..f842c75 --- /dev/null +++ b/Oganesson-TinyLlama-1.2B.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6dddf8c58701a9c6a487269e7917868de6c2fc184de94ec6432219d47c161db2 +size 807690496 diff --git a/README.md b/README.md new file mode 100644 index 0000000..90530e3 --- /dev/null +++ b/README.md @@ -0,0 +1,38 @@ +--- +license: apache-2.0 +language: +- en +base_model: +- prithivMLmods/Oganesson-TinyLlama-1.2B +pipeline_tag: text-generation +library_name: transformers +tags: +- text-generation-inference +- code +- math +- llama-3.2 +--- +# **Oganesson-TinyLlama-1.2B-GGUF** + +> **Oganesson-TinyLlama-1.2B** is a lightweight and efficient language model built on the **LLaMA 3.2 1.2B** architecture. Fine-tuned for **general-purpose inference**, **mathematical reasoning**, and **code generation**, it’s ideal for edge devices, personal assistants, and educational applications requiring a compact yet capable model. + +## Model File + +| File Name | Size | Format | +|-----------------------------------------------|---------|--------| +| Oganesson-TinyLlama-1.2B.BF16.gguf | 2.48 GB | BF16 | +| Oganesson-TinyLlama-1.2B.F16.gguf | 2.48 GB | F16 | +| Oganesson-TinyLlama-1.2B.F32.gguf | 4.95 GB | F32 | +| Oganesson-TinyLlama-1.2B.Q4_K_M.gguf | 808 MB | Q4_K_M | +| .gitattributes | 1.8 kB | - | +| README.md | 212 B | - | +| config.json | 31 B | JSON | + +## Quants Usage + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000..beb74f6 --- /dev/null +++ b/config.json @@ -0,0 +1,3 @@ +{ + "model_type": "llama" +} \ No newline at end of file diff --git a/configuration.json b/configuration.json new file mode 100644 index 0000000..159097f --- /dev/null +++ b/configuration.json @@ -0,0 +1 @@ +{"framework": "pytorch", "task": "others", "allow_remote": true} \ No newline at end of file