Tested at Q8_0 quantization. 
**Method:** * **Training Method:** DoRA (Weight-Decomposed LoRA) * **Target Modules** `all-linear` * **LoRA Rank:** 32 * **LoRA Alpha:** 32 * **LoRA Dropout:** 0.05 **Hyperparameters:** * **Batch Size:** 2 (Per-device) * **Gradient Accumulation:** 2 * **Epochs:** 2 * **Learning Rate:** 1e-4 * **Optimizer:** `adamw_torch_fused` * **LR Scheduler:** `cosine` * **Noise Level:** `neftune_noise_alpha=5`