Second Order
Second-order methods in machine learning leverage curvature information, primarily through Hessian matrices or their approximations, to improve optimization efficiency and model performance compared to first-order methods. Current research focuses on developing computationally tractable second-order algorithms, such as those employing diagonal Hessian approximations or low-rank matrix factorizations, for training large-scale models like LLMs and improving reinforcement learning. These advancements are significant because they offer faster convergence, enhanced generalization, and improved robustness in various applications, including image classification, natural language processing, and robotics.
Papers
November 17, 2022
November 11, 2022
November 3, 2022
October 26, 2022
October 20, 2022
October 18, 2022
October 11, 2022
October 10, 2022
October 6, 2022
September 19, 2022
September 6, 2022
September 2, 2022
September 1, 2022
August 23, 2022
August 3, 2022
July 30, 2022
July 19, 2022
July 17, 2022