Early Stage Convergence
Early stage convergence in machine learning focuses on understanding and improving the initial phases of training algorithms, aiming to accelerate convergence speed and enhance generalization performance. Current research investigates this through the lens of various optimization algorithms (e.g., Adam, SGD, FedProx), model architectures (e.g., transformers, diffusion models), and specific problem domains (e.g., federated learning, collaborative filtering). These studies leverage techniques from dynamical systems theory and optimal transport to establish convergence guarantees and bounds, ultimately contributing to more efficient and robust machine learning systems across diverse applications.
Papers
April 14, 2022
April 13, 2022
March 30, 2022
March 29, 2022
March 16, 2022
February 24, 2022
February 22, 2022
February 14, 2022
February 13, 2022
February 8, 2022
February 7, 2022
February 2, 2022
January 27, 2022
January 26, 2022
January 25, 2022
January 18, 2022
January 4, 2022