--- license: gemma language: - en base_model: kth8/gemma-3-1b-it-Conversation datasets: - kth8/multi-turn-conversation-50000x pipeline_tag: text-generation library_name: transformers tags: - sft - trl - unsloth - google - gemma - gemma3 - gemma3_text --- ![logo](https://storage.googleapis.com/gweb-developer-goog-blog-assets/images/gemma-3_2.original.png) A fine-tune of [unsloth/gemma-3-1b-it](https://huggingface.co/unsloth/gemma-3-1b-it) on the [kth8/multi-turn-conversation-50000x](https://huggingface.co/datasets/kth8/multi-turn-conversation-50000x) dataset. ## Usage example **System prompt** ``` You are a helpful assistant. ``` **User prompt** ``` Hey there! How's it going? ``` **Assistant response** ``` Hey! I'm doing great, thanks for asking! I'm here and ready to help with whatever you need. What's on your mind today? ``` ## Model Details - Base Model: `unsloth/gemma-3-1b-it` - Parameter Count: 999885952 - Precision: torch.bfloat16 ## Training Settings ### Hardware - GPU: NVIDIA RTX PRO 6000 Blackwell Server Edition ### PEFT - Rank: 32 - LoRA alpha: 64 - Modules: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj - Gradient checkpointing: unsloth ### SFT - Epoch: 2 - Batch size: 48 - Gradient Accumulation steps: 1 - Warmup ratio: 0.1 - Learning rate: 0.0002 - Optimizer: adamw_torch_fused - Learning rate scheduler: cosine ## Training stats - Global step: 1996 - Training runtime (seconds): 6834.1445 - Average training loss: 1.1743444665400442 - Final validation loss: 1.1191450357437134 ## Framework versions - Unsloth: 2026.3.8 - TRL: 0.22.2 - Transformers: 4.56.2 - Pytorch: 2.10.0+cu128 - Datasets: 4.8.3 - Tokenizers: 0.22.2 ## License This model is released under the Gemma license. See the [Gemma Terms of Use](https://ai.google.dev/gemma/terms) and [Prohibited Use Policy](https://policies.google.com/terms/generative-ai/use-policy) regarding the use of Gemma-generated content.