Stochastic Proximal Gradient

Stochastic proximal gradient methods extend stochastic gradient descent to handle optimization problems with non-smooth regularizers, crucial for many machine learning tasks. Current research focuses on improving convergence rates and sample complexity, particularly for non-convex problems, through techniques like variance reduction (e.g., using SPIDER or SARAH algorithms) and momentum methods (e.g., Polyak momentum). These advancements are significant because they enable efficient training of complex models in various applications, including reinforcement learning and robust optimization, by addressing challenges posed by noisy data and non-convex objective functions.

Papers