74 lines
1.9 KiB
Markdown
74 lines
1.9 KiB
Markdown
|
|
---
|
||
|
|
license: gemma
|
||
|
|
language:
|
||
|
|
- en
|
||
|
|
base_model: kth8/gemma-3-1b-it-Conversation
|
||
|
|
datasets:
|
||
|
|
- kth8/multi-turn-conversation-50000x
|
||
|
|
pipeline_tag: text-generation
|
||
|
|
library_name: transformers
|
||
|
|
tags:
|
||
|
|
- sft
|
||
|
|
- trl
|
||
|
|
- unsloth
|
||
|
|
- google
|
||
|
|
- gemma
|
||
|
|
- gemma3
|
||
|
|
- gemma3_text
|
||
|
|
---
|
||
|
|

|
||
|
|
A fine-tune of [unsloth/gemma-3-1b-it](https://huggingface.co/unsloth/gemma-3-1b-it) on the [kth8/multi-turn-conversation-50000x](https://huggingface.co/datasets/kth8/multi-turn-conversation-50000x) dataset.
|
||
|
|
|
||
|
|
## Usage example
|
||
|
|
**System prompt**
|
||
|
|
```
|
||
|
|
You are a helpful assistant.
|
||
|
|
```
|
||
|
|
**User prompt**
|
||
|
|
```
|
||
|
|
Hey there! How's it going?
|
||
|
|
```
|
||
|
|
**Assistant response**
|
||
|
|
```
|
||
|
|
Hey! I'm doing great, thanks for asking! I'm here and ready to help with whatever you need. What's on your mind today?
|
||
|
|
```
|
||
|
|
## Model Details
|
||
|
|
- Base Model: `unsloth/gemma-3-1b-it`
|
||
|
|
- Parameter Count: 999885952
|
||
|
|
- Precision: torch.bfloat16
|
||
|
|
|
||
|
|
## Training Settings
|
||
|
|
### Hardware
|
||
|
|
- GPU: NVIDIA RTX PRO 6000 Blackwell Server Edition
|
||
|
|
|
||
|
|
### PEFT
|
||
|
|
- Rank: 32
|
||
|
|
- LoRA alpha: 64
|
||
|
|
- Modules: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
|
||
|
|
- Gradient checkpointing: unsloth
|
||
|
|
|
||
|
|
### SFT
|
||
|
|
- Epoch: 2
|
||
|
|
- Batch size: 48
|
||
|
|
- Gradient Accumulation steps: 1
|
||
|
|
- Warmup ratio: 0.1
|
||
|
|
- Learning rate: 0.0002
|
||
|
|
- Optimizer: adamw_torch_fused
|
||
|
|
- Learning rate scheduler: cosine
|
||
|
|
|
||
|
|
## Training stats
|
||
|
|
- Global step: 1996
|
||
|
|
- Training runtime (seconds): 6834.1445
|
||
|
|
- Average training loss: 1.1743444665400442
|
||
|
|
- Final validation loss: 1.1191450357437134
|
||
|
|
|
||
|
|
## Framework versions
|
||
|
|
- Unsloth: 2026.3.8
|
||
|
|
- TRL: 0.22.2
|
||
|
|
- Transformers: 4.56.2
|
||
|
|
- Pytorch: 2.10.0+cu128
|
||
|
|
- Datasets: 4.8.3
|
||
|
|
- Tokenizers: 0.22.2
|
||
|
|
|
||
|
|
## License
|
||
|
|
This model is released under the Gemma license. See the [Gemma Terms of Use](https://ai.google.dev/gemma/terms) and [Prohibited Use Policy](https://policies.google.com/terms/generative-ai/use-policy) regarding the use of Gemma-generated content.
|