Infinite Width

Research on "infinite width" in neural networks investigates the theoretical properties of networks with infinitely many neurons per layer, aiming to simplify analysis and gain insights into the behavior of their finite-width counterparts. Current work focuses on understanding the dynamics of training with various optimizers (like Adam and SGD), exploring the role of feature learning and initialization strategies in different architectures (including MLPs and CNNs), and characterizing the resulting kernel functions. These analyses provide a more rigorous understanding of neural network training, potentially leading to improved optimization algorithms and hyperparameter tuning strategies for practical applications.

Papers