Learning Rate
Learning rate, a crucial hyperparameter in training neural networks, dictates the step size during optimization. Current research focuses on developing adaptive learning rate schedules, such as warmup-stable-decay and learning rate path switching, to improve training efficiency and generalization, particularly for large language models and other deep learning architectures. These advancements aim to address challenges like finding optimal learning rates across varying model sizes, datasets, and training durations, ultimately leading to faster convergence and better model performance. The impact extends to various applications, from natural language processing and computer vision to scientific computing and reinforcement learning.
Papers
Making Self-supervised Learning Robust to Spurious Correlation via Learning-speed Aware Sampling
Weicheng Zhu, Sheng Liu, Carlos Fernandez-Granda, Narges Razavian
Sensitivity-Based Layer Insertion for Residual and Feedforward Neural Networks
Evelyn Herberg, Roland Herzog, Frederik Köhne, Leonie Kreis, Anton Schiela
Model-free Posterior Sampling via Learning Rate Randomization
Daniil Tiapkin, Denis Belomestny, Daniele Calandriello, Eric Moulines, Remi Munos, Alexey Naumov, Pierre Perrault, Michal Valko, Pierre Menard
Closing the Gap Between the Upper Bound and the Lower Bound of Adam's Iteration Complexity
Bohan Wang, Jingwen Fu, Huishuai Zhang, Nanning Zheng, Wei Chen