Early Stage Convergence
Early stage convergence in machine learning focuses on understanding and improving the initial phases of training algorithms, aiming to accelerate convergence speed and enhance generalization performance. Current research investigates this through the lens of various optimization algorithms (e.g., Adam, SGD, FedProx), model architectures (e.g., transformers, diffusion models), and specific problem domains (e.g., federated learning, collaborative filtering). These studies leverage techniques from dynamical systems theory and optimal transport to establish convergence guarantees and bounds, ultimately contributing to more efficient and robust machine learning systems across diverse applications.
Papers
February 6, 2024
February 1, 2024
January 28, 2024
January 17, 2024
January 6, 2024
December 28, 2023
December 27, 2023
December 21, 2023
December 19, 2023
December 14, 2023
December 11, 2023
December 10, 2023
December 6, 2023
November 30, 2023
November 21, 2023
November 17, 2023
November 7, 2023
November 2, 2023