Files
gemma-7b-ultrachat-sft/.ipynb_checkpoints/README-checkpoint.md
ModelHub XC c403cebb05 初始化项目,由ModelHub XC社区提供模型
Model: CorticalStack/gemma-7b-ultrachat-sft
Source: Original Platform
2026-04-16 14:56:04 +08:00

843 B

license
license
apache-2.0

gemma-7b-ultrachat-sft

gemma-7b-ultrachat-sft is an SFT fine-tuned version of google/gemma-7b using the stingning/ultrachat dataset.

Fine-tuning configuration

LoRA

  • r: 8
  • LoRA alpha: 16
  • LoRA dropout: 0.1

Training arguments

  • Epochs: 1
  • Batch size: 4
  • Gradient accumulation steps: 6
  • Optimizer: paged_adamw_32bit
  • Max steps: 100
  • Learning rate: 0.0002
  • Weight decay: 0.001
  • Learning rate scheduler type: constant
  • Max seq length: 2048

Trained with Unsloth and Huggingface's TRL library.