Knowledge Unlearning
Knowledge unlearning aims to selectively remove specific information from machine learning models, particularly large language models (LLMs), without requiring complete retraining. Current research focuses on developing effective algorithms for this task, often employing techniques like gradient-based methods, knowledge distillation, and model inversion, applied to various architectures including LLMs and graph neural networks. This field is crucial for addressing privacy concerns, complying with data protection regulations (like GDPR), and enhancing the safety and trustworthiness of AI systems by mitigating the risk of sensitive data leakage or harmful outputs.
Papers
Towards Efficient Target-Level Machine Unlearning Based on Essential Graph
Heng Xu, Tianqing Zhu, Lefeng Zhang, Wanlei Zhou, Wei Zhao
RWKU: Benchmarking Real-World Knowledge Unlearning for Large Language Models
Zhuoran Jin, Pengfei Cao, Chenhao Wang, Zhitao He, Hongbang Yuan, Jiachun Li, Yubo Chen, Kang Liu, Jun Zhao