Gradient Descent Trajectory

Gradient descent trajectory analysis focuses on understanding the path taken by optimization algorithms, like stochastic gradient descent (SGD) and Adam, as they search for optimal model parameters. Current research investigates the trajectory's relationship to the loss function's landscape, particularly focusing on how the trajectory interacts with the Hessian and gradient matrices, especially in high-dimensional settings and with models such as neural networks (including ResNets). These studies aim to explain phenomena like implicit regularization and the "edge of stability," ultimately seeking to improve algorithm design and convergence guarantees for various machine learning tasks. This work has implications for both theoretical understanding of optimization and practical improvements in training efficiency and generalization performance.

Papers