Second Order Optimization
Second-order optimization methods aim to accelerate the training of machine learning models by incorporating information about the curvature of the loss function, leading to faster convergence compared to first-order methods. Current research focuses on developing efficient approximations of the Hessian matrix, crucial for scalability, with algorithms like K-FAC and Shampoo being prominent, as well as exploring hybrid approaches combining first and second-order techniques. These advancements are significant because they improve the efficiency and effectiveness of training large-scale models across diverse applications, including deep learning, reinforcement learning, and scientific machine learning.
Papers
November 12, 2024
October 29, 2024
October 18, 2024
August 30, 2024
June 5, 2024
May 16, 2024
April 28, 2024
March 8, 2024
January 12, 2024
December 19, 2023
October 18, 2023
August 4, 2023
February 16, 2023
December 1, 2022
October 20, 2022
June 23, 2022
March 29, 2022
March 18, 2022