Early Stage Convergence
Early stage convergence in machine learning focuses on understanding and improving the initial phases of training algorithms, aiming to accelerate convergence speed and enhance generalization performance. Current research investigates this through the lens of various optimization algorithms (e.g., Adam, SGD, FedProx), model architectures (e.g., transformers, diffusion models), and specific problem domains (e.g., federated learning, collaborative filtering). These studies leverage techniques from dynamical systems theory and optimal transport to establish convergence guarantees and bounds, ultimately contributing to more efficient and robust machine learning systems across diverse applications.
Papers
September 23, 2023
September 21, 2023
September 15, 2023
September 14, 2023
September 12, 2023
September 11, 2023
September 3, 2023
September 1, 2023
August 16, 2023
August 7, 2023
July 30, 2023
July 26, 2023
July 23, 2023
July 21, 2023
July 20, 2023
July 10, 2023
July 5, 2023
June 23, 2023