Standard Gradient Descent
Standard gradient descent (SGD) is a fundamental optimization algorithm used to train machine learning models, aiming to minimize a loss function by iteratively adjusting model parameters. Current research focuses on improving SGD's efficiency and robustness, exploring variations like adaptive batch size and step size methods, and incorporating regularization techniques to enhance generalization and mitigate issues like overfitting and sensitivity to noisy data. These advancements are crucial for training increasingly complex models on large datasets, impacting diverse applications from image recognition to solving complex scientific problems.
Papers
October 17, 2024
September 25, 2024
July 22, 2024
February 7, 2024
March 31, 2023
January 31, 2023
November 22, 2022
October 7, 2022
October 6, 2022
July 24, 2022
July 16, 2022
July 13, 2022
June 13, 2022
March 3, 2022
February 26, 2022
February 13, 2022
January 25, 2022
December 4, 2021