Files
Llama-3.1-8B-Instruct_SDFT_…/README.md
ModelHub XC 0a3cd7b7bc 初始化项目,由ModelHub XC社区提供模型
Model: Neelectric/Llama-3.1-8B-Instruct_SDFT_mathv00.06
Source: Original Platform
2026-04-10 15:44:08 +08:00

2.5 KiB

base_model, datasets, library_name, model_name, tags, licence
base_model datasets library_name model_name tags licence
meta-llama/Llama-3.1-8B-Instruct Neelectric/OpenR1-Math-220k_all_SDFT_nr transformers Llama-3.1-8B-Instruct_SDFT_mathv00.06
generated_from_trainer
open-r1
trl
sdft
license

Model Card for Llama-3.1-8B-Instruct_SDFT_mathv00.06

This model is a fine-tuned version of meta-llama/Llama-3.1-8B-Instruct on the Neelectric/OpenR1-Math-220k_all_SDFT_nr dataset. It has been trained using TRL.

Quick start

from transformers import pipeline

question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="Neelectric/Llama-3.1-8B-Instruct_SDFT_mathv00.06", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])

Training procedure

Visualize in Weights & Biases

This model was trained with SDFT, a method introduced in Self-Training with On-Policy Self-Distillation for Language Model Alignment.

Framework versions

  • TRL: 1.1.0.dev0
  • Transformers: 4.57.6
  • Pytorch: 2.9.0
  • Datasets: 4.8.4
  • Tokenizers: 0.22.2

Citations

Cite SDFT as:

@article{hubotter2026selftraining,
    title        = {{Self-Training with On-Policy Self-Distillation for Language Model Alignment}},
    author       = {Jonas H\"ubotter and Frederike L\"ubeck and Lejs Behric and Anton Baumann and Marco Bagatella and Daniel Marta and Ido Hakimi and Idan Shenfeld and Thomas Kleine Buening and Carlos Guestrin and Andreas Krause},
    year         = 2026,
    eprint       = {arXiv:2601.19897}
}

Cite TRL as:

@software{vonwerra2020trl,
  title   = {{TRL: Transformers Reinforcement Learning}},
  author  = {von Werra, Leandro and Belkada, Younes and Tunstall, Lewis and Beeching, Edward and Thrush, Tristan and Lambert, Nathan and Huang, Shengyi and Rasul, Kashif and Gallouédec, Quentin},
  license = {Apache-2.0},
  url     = {https://github.com/huggingface/trl},
  year    = {2020}
}