Natural Gradient
Natural gradients are a powerful tool in optimization, aiming to improve the efficiency and stability of training complex models by accounting for the underlying geometry of the parameter space. Current research focuses on applying natural gradient methods to diverse areas, including distributed learning (e.g., through gradient compression and efficient client selection), inverse problems (using diffusion models), and neural network training (e.g., via regularization and novel optimizers like DiffGrad and AdEMAMix). These advancements have significant implications for improving the performance and robustness of machine learning models across various applications, from image processing and medical image analysis to scientific computing and federated learning.
Papers
Random orthogonal additive filters: a solution to the vanishing/exploding gradient of deep neural networks
Andrea Ceni
Alternating Differentiation for Optimization Layers
Haixiang Sun, Ye Shi, Jingya Wang, Hoang Duong Tuan, H. Vincent Poor, Dacheng Tao
Limitations of neural network training due to numerical instability of backpropagation
Clemens Karner, Vladimir Kazeev, Philipp Christian Petersen