Infinite Width
Research on "infinite width" in neural networks investigates the theoretical properties of networks with infinitely many neurons per layer, aiming to simplify analysis and gain insights into the behavior of their finite-width counterparts. Current work focuses on understanding the dynamics of training with various optimizers (like Adam and SGD), exploring the role of feature learning and initialization strategies in different architectures (including MLPs and CNNs), and characterizing the resulting kernel functions. These analyses provide a more rigorous understanding of neural network training, potentially leading to improved optimization algorithms and hyperparameter tuning strategies for practical applications.
Papers
November 4, 2024
June 27, 2024
December 19, 2023
December 6, 2023
September 4, 2023
August 3, 2023
April 6, 2023
November 29, 2022
October 3, 2022
May 24, 2022
May 17, 2022
February 24, 2022
February 1, 2022