From 43f2175bba68c193cd948617d3e998bbc43a0ba9 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sun, 12 Apr 2026 12:55:58 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: kth8/gemma-3-1b-it-Conversation-GGUF Source: Original Platform --- .gitattributes | 40 ++++++++++++++ README.md | 74 ++++++++++++++++++++++++++ gemma-3-1b-it-Conversation-Q4_0.gguf | 3 ++ gemma-3-1b-it-Conversation-Q4_K_M.gguf | 3 ++ gemma-3-1b-it-Conversation-Q6_K.gguf | 3 ++ gemma-3-1b-it-Conversation-Q8_0.gguf | 3 ++ gemma-3-1b-it-Conversation-bf16.gguf | 3 ++ 7 files changed, 129 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 gemma-3-1b-it-Conversation-Q4_0.gguf create mode 100644 gemma-3-1b-it-Conversation-Q4_K_M.gguf create mode 100644 gemma-3-1b-it-Conversation-Q6_K.gguf create mode 100644 gemma-3-1b-it-Conversation-Q8_0.gguf create mode 100644 gemma-3-1b-it-Conversation-bf16.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..13778e5 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,40 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-Conversation-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-Conversation-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-Conversation-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-Conversation-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-Conversation-bf16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..84da0db --- /dev/null +++ b/README.md @@ -0,0 +1,74 @@ +--- +license: gemma +language: +- en +base_model: kth8/gemma-3-1b-it-Conversation +datasets: +- kth8/multi-turn-conversation-50000x +pipeline_tag: text-generation +library_name: transformers +tags: +- sft +- trl +- unsloth +- google +- gemma +- gemma3 +- gemma3_text +--- +![logo](https://storage.googleapis.com/gweb-developer-goog-blog-assets/images/gemma-3_2.original.png) +A fine-tune of [unsloth/gemma-3-1b-it](https://huggingface.co/unsloth/gemma-3-1b-it) on the [kth8/multi-turn-conversation-50000x](https://huggingface.co/datasets/kth8/multi-turn-conversation-50000x) dataset. + +## Usage example +**System prompt** +``` +You are a helpful assistant. +``` +**User prompt** +``` +Hey there! How's it going? +``` +**Assistant response** +``` +Hey! I'm doing great, thanks for asking! I'm here and ready to help with whatever you need. What's on your mind today? +``` +## Model Details +- Base Model: `unsloth/gemma-3-1b-it` +- Parameter Count: 999885952 +- Precision: torch.bfloat16 + +## Training Settings +### Hardware +- GPU: NVIDIA RTX PRO 6000 Blackwell Server Edition + +### PEFT +- Rank: 32 +- LoRA alpha: 64 +- Modules: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj +- Gradient checkpointing: unsloth + +### SFT +- Epoch: 2 +- Batch size: 48 +- Gradient Accumulation steps: 1 +- Warmup ratio: 0.1 +- Learning rate: 0.0002 +- Optimizer: adamw_torch_fused +- Learning rate scheduler: cosine + +## Training stats +- Global step: 1996 +- Training runtime (seconds): 6834.1445 +- Average training loss: 1.1743444665400442 +- Final validation loss: 1.1191450357437134 + +## Framework versions +- Unsloth: 2026.3.8 +- TRL: 0.22.2 +- Transformers: 4.56.2 +- Pytorch: 2.10.0+cu128 +- Datasets: 4.8.3 +- Tokenizers: 0.22.2 + +## License +This model is released under the Gemma license. See the [Gemma Terms of Use](https://ai.google.dev/gemma/terms) and [Prohibited Use Policy](https://policies.google.com/terms/generative-ai/use-policy) regarding the use of Gemma-generated content. \ No newline at end of file diff --git a/gemma-3-1b-it-Conversation-Q4_0.gguf b/gemma-3-1b-it-Conversation-Q4_0.gguf new file mode 100644 index 0000000..c488273 --- /dev/null +++ b/gemma-3-1b-it-Conversation-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0660f9eaa1ee0eb1d363ef16a4601c8a9bed4d71b8647693855b83a5eb639c7a +size 720425952 diff --git a/gemma-3-1b-it-Conversation-Q4_K_M.gguf b/gemma-3-1b-it-Conversation-Q4_K_M.gguf new file mode 100644 index 0000000..ae2cea6 --- /dev/null +++ b/gemma-3-1b-it-Conversation-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:acba3cfbdc674871d6e8996ca2a25aa989d473bcc303720eb25ea7d21000fd3d +size 806058720 diff --git a/gemma-3-1b-it-Conversation-Q6_K.gguf b/gemma-3-1b-it-Conversation-Q6_K.gguf new file mode 100644 index 0000000..193a3e3 --- /dev/null +++ b/gemma-3-1b-it-Conversation-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ad43a3d664d5c6e3a2abbac7aa70609e1b48b8f5428d62d40734052958f924e +size 1011739104 diff --git a/gemma-3-1b-it-Conversation-Q8_0.gguf b/gemma-3-1b-it-Conversation-Q8_0.gguf new file mode 100644 index 0000000..45b3cf7 --- /dev/null +++ b/gemma-3-1b-it-Conversation-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7f6557a8df35a9717ea2ce5131874ad3cc924347188d3b9c9a0d925256a3025 +size 1069306848 diff --git a/gemma-3-1b-it-Conversation-bf16.gguf b/gemma-3-1b-it-Conversation-bf16.gguf new file mode 100644 index 0000000..f27b7e5 --- /dev/null +++ b/gemma-3-1b-it-Conversation-bf16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:209e6ba021a09172e50e7f93e0657df5e066cc08605409e281396c3eaf909daa +size 2006574048