Gradient Norm
Gradient norm, the magnitude of the gradient vector in optimization algorithms, is a central concept in deep learning research, with current efforts focusing on understanding its role in algorithm convergence, model robustness, and efficient training. Research investigates the impact of gradient norm on various optimization algorithms (e.g., Adam, SGD, RMSProp) and its relationship to model generalization and adversarial robustness, often within the context of specific architectures like vision transformers. Understanding and controlling gradient norm is crucial for improving the efficiency, stability, and reliability of deep learning models across diverse applications, from image classification to federated learning.
Papers
October 3, 2022
September 21, 2022
September 19, 2022
September 2, 2022
August 23, 2022
August 21, 2022
August 11, 2022
June 29, 2022
June 27, 2022
June 23, 2022
June 7, 2022
June 5, 2022
May 30, 2022
May 26, 2022
May 4, 2022
March 24, 2022
February 14, 2022
February 11, 2022
February 9, 2022