Continued pretraining for Macedonian language on lvstck/macedonian-corpus-cleaned-dedup.
Trained using LoRA adapters on a single A100.
Training
Stage 1: 0.4 epochs on Google Colab A100
Stage 2: 0.9 epochs on A100 - FINKI GPU cluster
Total: 1.3 epochs
Training Metrics
TensorBoard logs are available in the Training metrics tab of this model repository. Logs only available for STAGE 2
The logs include training loss, learning rate, gradient norm, and evaluation metrics .