Wide Neural Network

Wide neural networks, characterized by a large number of neurons in their layers, are a focus of intense research due to their unique theoretical properties and practical performance. Current research investigates their training dynamics, generalization capabilities, and connections to Gaussian processes and kernel methods, often focusing on architectures like ResNets and deep equilibrium models, and employing techniques like neural tangent kernel analysis. These studies aim to provide a deeper understanding of how and why these networks generalize well, even when overparameterized, leading to improved model design and training strategies with implications for various applications, including image classification and solving partial differential equations.

Papers