Figure 1: Schematic overview of unlearning trace detection. |
This is the official code repository for the ICLR 2026 paper Unlearning Isn't Invisible: Detecting Unlearning Traces in LLMs from Model Outputs.
- 🎉 [Jan.26.2026] Our paper is accepted at ICLR 2026!
- 🏆 [Jun.10.2025] Our paper’s short version accepted for Oral at MUGen@ICML’25!
- 🔥 Check out our related ICLR 2026 paper: Safety Mirage, which proposes machine unlearning as a more robust alignment alternative for VLM safety fine-tuning.
Please see Data.md.
Please see Unlearn.md.
Please see Installation.md.
Please see Response.md
Please see Classification.md
If you find out our paper or code helpful, please cite our work~
@article{chen2025unlearning,
title={Unlearning Isn't Invisible: Detecting Unlearning Traces in LLMs from Model Outputs},
author={Chen, Yiwei and Pal, Soumyadeep and Zhang, Yimeng and Qu, Qing and Liu, Sijia},
journal={arXiv preprint arXiv:2506.14003},
year={2025}
}
