Unlearning Framework
Machine unlearning aims to remove the influence of specific data points from trained machine learning models, addressing privacy concerns and promoting data control. Current research focuses on developing efficient unlearning algorithms for various model architectures, including deep neural networks, large language models, and graph neural networks, often employing techniques like gradient manipulation, parameter editing, and contrastive learning. This field is crucial for ensuring compliance with data privacy regulations and enhancing the trustworthiness and safety of deployed AI systems across diverse applications, from medical diagnosis to personalized recommendations. Challenges remain in balancing effective unlearning with the preservation of model utility and robustness against adversarial attacks.
Papers
PISTOL: Dataset Compilation Pipeline for Structural Unlearning of LLMs
Xinchi Qiu, William F. Shen, Yihong Chen, Nicola Cancedda, Pontus Stenetorp, Nicholas D. Lane
Towards Scalable Exact Machine Unlearning Using Parameter-Efficient Fine-Tuning
Somnath Basu Roy Chowdhury, Krzysztof Choromanski, Arijit Sehanobish, Avinava Dubey, Snigdha Chaturvedi
Machine Unlearning with Minimal Gradient Dependence for High Unlearning Ratios
Tao Huang, Ziyang Chen, Jiayang Meng, Qingyu Huang, Xu Yang, Xun Yi, Ibrahim Khalil
Towards Efficient Target-Level Machine Unlearning Based on Essential Graph
Heng Xu, Tianqing Zhu, Lefeng Zhang, Wanlei Zhou, Wei Zhao
RWKU: Benchmarking Real-World Knowledge Unlearning for Large Language Models
Zhuoran Jin, Pengfei Cao, Chenhao Wang, Zhitao He, Hongbang Yuan, Jiachun Li, Yubo Chen, Kang Liu, Jun Zhao