Gradient Norm
Gradient norm, the magnitude of the gradient vector in optimization algorithms, is a central concept in deep learning research, with current efforts focusing on understanding its role in algorithm convergence, model robustness, and efficient training. Research investigates the impact of gradient norm on various optimization algorithms (e.g., Adam, SGD, RMSProp) and its relationship to model generalization and adversarial robustness, often within the context of specific architectures like vision transformers. Understanding and controlling gradient norm is crucial for improving the efficiency, stability, and reliability of deep learning models across diverse applications, from image classification to federated learning.
101papers
Papers
March 29, 2025
Towards Understanding the Optimization Mechanisms in Deep Learning
Binchuan Qi, Wei Gong, Li LiTongji University●National Key Laboratory of Autonomous Intelligent Unmanned Systems●Shanghai Research Institute for Intelligent Autonomous...+1DC-SGD: Differentially Private SGD with Dynamic Clipping through Gradient Norm Distribution Estimation
Chengkun Wei, Weixian Li, Gong Chen, Wenzhi ChenZhejiang University●Ant Group●University of Virginia
March 18, 2025
March 14, 2025
February 3, 2025
February 2, 2025
January 31, 2025
January 15, 2025
November 24, 2024
October 30, 2024