Gradient Descent Optimizers
Gradient descent optimizers are algorithms used to train machine learning models by iteratively adjusting parameters to minimize a loss function. Current research focuses on improving their efficiency and robustness, exploring variations like adaptive learning rate methods (e.g., Adam, AdaGrad) and techniques to mitigate sensitivity to hyperparameter initialization (e.g., ActiveLR). These advancements aim to accelerate training, enhance model generalization, and reduce computational costs, impacting various fields from computer vision (using CNNs) to quantum machine learning.
Papers
July 17, 2024
March 22, 2023
January 24, 2023
November 9, 2022
July 19, 2022
March 5, 2022
January 18, 2022