Catastrophic Forgetting
Catastrophic forgetting describes the phenomenon where artificial neural networks, upon learning new tasks, lose previously acquired knowledge. Current research focuses on mitigating this issue through various strategies, including parameter-efficient fine-tuning methods (like LoRA), generative model-based data replay, and novel optimization algorithms that constrain gradient updates or leverage hierarchical task structures. Addressing catastrophic forgetting is crucial for developing robust and adaptable AI systems capable of continuous learning in real-world applications, particularly in domains like medical imaging, robotics, and natural language processing where data streams are constantly evolving.
Papers
Neuro-mimetic Task-free Unsupervised Online Learning with Continual Self-Organizing Maps
Hitesh Vaidya, Travis Desell, Ankur Mali, Alexander Ororbia
Bayesian Parameter-Efficient Fine-Tuning for Overcoming Catastrophic Forgetting
Haolin Chen, Philip N. Garner
Model Tailor: Mitigating Catastrophic Forgetting in Multi-modal Large Language Models
Didi Zhu, Zhongyi Sun, Zexi Li, Tao Shen, Ke Yan, Shouhong Ding, Kun Kuang, Chao Wu
Quantified Task Misalignment to Inform PEFT: An Exploration of Domain Generalization and Catastrophic Forgetting in CLIP
Laura Niss, Kevin Vogt-Lowell, Theodoros Tsiligkaridis
Layerwise Proximal Replay: A Proximal Point Method for Online Continual Learning
Jason Yoo, Yunpeng Liu, Frank Wood, Geoff Pleiss
INCPrompt: Task-Aware incremental Prompting for Rehearsal-Free Class-incremental Learning
Zhiyuan Wang, Xiaoyang Qu, Jing Xiao, Bokui Chen, Jianzong Wang
P2DT: Mitigating Forgetting in task-incremental Learning with progressive prompt Decision Transformer
Zhiyuan Wang, Xiaoyang Qu, Jing Xiao, Bokui Chen, Jianzong Wang
Beyond Anti-Forgetting: Multimodal Continual Instruction Tuning with Positive Forward Transfer
Junhao Zheng, Qianli Ma, Zhen Liu, Binquan Wu, Huawen Feng
Towards Continual Learning Desiderata via HSIC-Bottleneck Orthogonalization and Equiangular Embedding
Depeng Li, Tianqi Wang, Junwei Chen, Qining Ren, Kenji Kawaguchi, Zhigang Zeng
CEL: A Continual Learning Model for Disease Outbreak Prediction by Leveraging Domain Adaptation via Elastic Weight Consolidation
Saba Aslam, Abdur Rasool, Hongyan Wu, Xiaoli Li