Gradient Stability
Gradient stability in neural networks is a crucial area of research focusing on mitigating unstable gradients that hinder training and performance. Current efforts concentrate on developing techniques to stabilize gradients during training, including novel estimators, architectural modifications like low-curvature networks and orthogonal weight initialization, and improved training strategies such as temporal unrolling and contrastive loss functions. Addressing gradient instability is vital for improving the robustness, efficiency, and interpretability of neural networks across diverse applications, from image generation and physics simulation to binary neural network compression.
Papers
October 15, 2024
June 25, 2024
February 20, 2024
August 13, 2023
October 27, 2022
June 14, 2022