--- license: mit language: - en pipeline_tag: text-generation arxiv: - https://arxiv.org/abs/2508.06595 library_name: transformers --- ## Model Details Best [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) checkpoint unlearned using [RMU](https://arxiv.org/abs/2403.03218) with the Keyword-Bio forget set. For more details, please check [our paper](https://arxiv.org/abs/2508.06595). ### sources - Base model: [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) - Repository: [https://github.com/xyzhu123/Synthetic_Textbook) ### Performance | | WMDP-Bio | tinyMMLU | GSM8k | TriviaQA | |----------------------------------------|:---------:|:----------:|:-------:|:--------:| | Llama-3-8B-Instruct | 71.01 | 59.21 | 75.28 | 51.09 | | Llama-3-8B-Instruct_RMU_Keyword-Bio | 70.30 | 60.42 | 75.97 | 51.31 | ## Citation If you find this useful in your research, please consider citing our paper: ``` @misc{zhu2025llmunlearningexpertcurated, title={LLM Unlearning Without an Expert Curated Dataset}, author={Xiaoyuan Zhu and Muru Zhang and Ollie Liu and Robin Jia and Willie Neiswanger}, year={2025}, eprint={2508.06595}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2508.06595}, } ```