Shallow ReLU
Shallow ReLU networks, characterized by a single hidden layer using the Rectified Linear Unit (ReLU) activation function, are a focus of research aiming to understand their approximation capabilities and optimization landscapes. Current investigations explore their performance in various applications, including function approximation, control systems, and image processing, often focusing on the impact of random weight initialization and the development of efficient training algorithms like structure-guided Gauss-Newton methods. These studies are significant because they provide theoretical foundations for understanding the behavior of these simple yet powerful networks, leading to improved training strategies and a deeper understanding of their generalization properties.
Papers
Weighted variation spaces and approximation by shallow ReLU networks
Ronald DeVore, Robert D. Nowak, Rahul Parhi, Jonathan W. Siegel
Noisy Interpolation Learning with Shallow Univariate ReLU Networks
Nirmit Joshi, Gal Vardi, Nathan Srebro
Optimal Approximation of Zonoids and Uniform Approximation by Shallow Neural Networks
Jonathan W. Siegel