Continual Training
Continual training aims to enable machine learning models, particularly large language and vision models, to adapt to new data streams without catastrophic forgetting of previously learned information. Current research focuses on developing efficient algorithms and architectures, such as parameter-efficient fine-tuning methods and replay strategies, to address this challenge across various model types, including transformers and recurrent neural networks. This field is crucial for developing more sustainable and adaptable AI systems, improving their performance in dynamic real-world environments and reducing the environmental impact of frequent retraining.
Papers
March 12, 2023
March 11, 2023
February 7, 2023
October 13, 2022
October 11, 2022
September 30, 2022
May 24, 2022
December 31, 2021