license, language, base_model, pipeline_tag, library_name, tags, datasets
license language base_model pipeline_tag library_name tags datasets
mit
en
muse-bench/MUSE-news_target
text-generation transformers
unlearn
machine-unlearning
llm-unlearning
data-privacy
large-language-models
trustworthy-ai
trustworthy-machine-learning
language-model
muse-bench/MUSE-News

NPO-Unlearned w/ SAM Model on Task "MUSE NEWS"

Model Details

Loading the Model

import torch
from transformers import AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("OPTML-Group/NPO-SAM-MUSE-NEWS", torch_dtype=torch.bfloat16, trust_remote_code=True)

Citation

If you use this model in your research, please cite:

@article{fan2025towards,
  title={Towards LLM Unlearning Resilient to Relearning Attacks: A Sharpness-Aware Minimization Perspective and Beyond},
  author={Fan, Chongyu and Jia, Jinghan and Zhang, Yihua and Ramakrishna, Anil and Hong, Mingyi and Liu, Sijia},
  journal={arXiv preprint arXiv:2502.05374},
  year={2025}
}

Reporting Issues

Reporting issues with the model: github.com/OPTML-Group/Unlearn-Smooth

Description
Model synced from source: OPTML-Group/NPO-SAM-MUSE-NEWS
Readme 581 KiB