Two Layer Network
Two-layer networks, a fundamental building block in deep learning and network optimization, are the subject of intense research focused on understanding their training dynamics and generalization capabilities. Current studies explore various aspects, including gradient descent algorithms (e.g., large stepsize GD, multi-pass GD), the impact of activation functions and network width, and the role of outliers and simplicity bias in model behavior. These investigations aim to improve training efficiency, enhance generalization performance, and provide a deeper theoretical understanding of neural network learning, with implications for both machine learning algorithms and network optimization in diverse applications.
Papers
November 5, 2024
October 31, 2024
October 16, 2024
October 11, 2024
August 16, 2024
August 15, 2024
June 12, 2024
May 27, 2024
March 1, 2024
February 7, 2024
February 5, 2024
November 7, 2023
September 30, 2023
May 29, 2023
February 12, 2023
November 17, 2022
August 8, 2022
August 4, 2022