Large Learning Rate
Large learning rates in neural network training are a focus of current research, aiming to understand their impact on optimization dynamics and generalization performance. Studies explore how large learning rates affect loss landscapes, influencing training trajectories and the types of solutions found, with investigations spanning various architectures and optimizers including SGD, Adam, and their variants. This research is significant because it challenges conventional wisdom about optimal learning rate strategies and offers potential for more efficient and effective training of large-scale models, particularly in resource-constrained environments.
Papers
October 29, 2024
October 13, 2024
October 7, 2024
October 5, 2024
June 24, 2024
June 7, 2024
May 25, 2024
February 20, 2024
February 12, 2024
January 21, 2024
November 19, 2023
October 26, 2023
September 12, 2023
July 22, 2023
June 29, 2023
June 7, 2023
May 30, 2023
May 28, 2023