Two Layer Neural Network
Two-layer neural networks serve as a fundamental model for understanding the behavior of deeper networks, with research focusing on their optimization dynamics, generalization capabilities, and feature learning properties. Current investigations utilize stochastic gradient descent and related algorithms, often within the context of the neural tangent kernel approximation, to analyze convergence rates and the impact of hyperparameters like learning rate and network width. These studies provide crucial insights into the theoretical foundations of deep learning, informing the design of more efficient and robust algorithms and offering a clearer understanding of phenomena like spectral bias and the emergence of skills during training.
Papers
February 6, 2023
December 7, 2022
November 17, 2022
November 16, 2022
October 28, 2022
September 29, 2022
September 27, 2022
August 10, 2022
June 14, 2022
June 10, 2022
June 2, 2022
May 19, 2022
May 17, 2022
May 3, 2022
April 24, 2022
April 8, 2022
March 22, 2022
February 17, 2022
February 16, 2022