Paper ID: 2306.06613
Parameter-free version of Adaptive Gradient Methods for Strongly-Convex Functions
Deepak Gouda, Hassan Naveed, Salil Kamath
The optimal learning rate for adaptive gradient methods applied to {\lambda}-strongly convex functions relies on the parameters {\lambda} and learning rate {\eta}. In this paper, we adapt a universal algorithm along the lines of Metagrad, to get rid of this dependence on {\lambda} and {\eta}. The main idea is to concurrently run multiple experts and combine their predictions to a master algorithm. This master enjoys O(d log T) regret bounds.
Submitted: Jun 11, 2023