WEight AVERaging
Weight averaging (WA) is a technique used to improve the performance and generalization of deep learning models by combining the weights of multiple models trained independently or in parallel. Current research focuses on applying WA in diverse contexts, including domain generalization, continual learning, and adversarial training, often employing it within ensemble methods or as a component of novel optimization algorithms like stochastic weight averaging (SWA) and its variants. The effectiveness of WA stems from its ability to mitigate overfitting, enhance robustness, and accelerate training, leading to improved accuracy and efficiency across various applications, particularly in computer vision and natural language processing.
Papers
November 2, 2024
June 21, 2024
May 27, 2024
January 31, 2024
October 20, 2023
October 5, 2023
October 3, 2023
September 20, 2023
August 22, 2023
June 13, 2023
June 5, 2023
April 23, 2023
April 6, 2023
March 14, 2023
October 27, 2022
September 29, 2022
June 30, 2022
May 19, 2022
April 26, 2022
February 21, 2022