Preconditioned Gradient Descent
Preconditioned gradient descent (PGD) enhances the efficiency and effectiveness of gradient-based optimization methods by incorporating problem-specific information to guide the search for optimal solutions. Current research focuses on adapting PGD to various contexts, including overparameterized neural networks, Wasserstein spaces, and stochastic optimization settings, often employing techniques like Kronecker factorization to manage computational costs. These advancements aim to improve convergence rates, reduce hyperparameter sensitivity, and enable the solution of large-scale optimization problems arising in machine learning and other fields, ultimately leading to more efficient and robust algorithms.
Papers
July 16, 2024
June 13, 2024
February 7, 2024
November 28, 2023
September 5, 2023
May 30, 2023
June 7, 2022