Gradient Norm
Gradient norm, the magnitude of the gradient vector in optimization algorithms, is a central concept in deep learning research, with current efforts focusing on understanding its role in algorithm convergence, model robustness, and efficient training. Research investigates the impact of gradient norm on various optimization algorithms (e.g., Adam, SGD, RMSProp) and its relationship to model generalization and adversarial robustness, often within the context of specific architectures like vision transformers. Understanding and controlling gradient norm is crucial for improving the efficiency, stability, and reliability of deep learning models across diverse applications, from image classification to federated learning.
Papers
October 26, 2023
October 16, 2023
October 11, 2023
September 29, 2023
September 15, 2023
August 23, 2023
July 16, 2023
July 7, 2023
June 14, 2023
June 8, 2023
June 2, 2023
May 29, 2023
May 21, 2023
May 18, 2023
May 2, 2023
April 27, 2023
April 17, 2023
April 4, 2023
April 3, 2023