Gradient Norm
Gradient norm, the magnitude of the gradient vector in optimization algorithms, is a central concept in deep learning research, with current efforts focusing on understanding its role in algorithm convergence, model robustness, and efficient training. Research investigates the impact of gradient norm on various optimization algorithms (e.g., Adam, SGD, RMSProp) and its relationship to model generalization and adversarial robustness, often within the context of specific architectures like vision transformers. Understanding and controlling gradient norm is crucial for improving the efficiency, stability, and reliability of deep learning models across diverse applications, from image classification to federated learning.
Papers
March 22, 2024
March 18, 2024
March 7, 2024
March 5, 2024
March 3, 2024
February 21, 2024
February 4, 2024
January 29, 2024
January 16, 2024
January 9, 2024
December 21, 2023
December 20, 2023
December 19, 2023
November 12, 2023
November 10, 2023
November 5, 2023
November 3, 2023