Generalization Error
Generalization error, the difference between a model's performance on training and unseen data, is a central challenge in machine learning. Current research focuses on understanding and mitigating this error across various model architectures, including linear models, neural networks (especially deep and overparameterized ones), and graph neural networks, often employing techniques like stochastic gradient descent, early stopping, and ensemble methods such as bagging. This research aims to develop tighter theoretical bounds on generalization error and improve model selection and assessment, particularly under conditions like data scarcity, distribution shifts, and adversarial attacks. Improved understanding of generalization error is crucial for building more reliable and robust machine learning systems across diverse applications.
Papers
Gradient flow in the gaussian covariate model: exact solution of learning curves and multiple descent structures
Antoine Bodin, Nicolas Macris
A Statistical Model for Predicting Generalization in Few-Shot Classification
Yassir Bendou, Vincent Gripon, Bastien Pasdeloup, Lukas Mauch, Stefan Uhlich, Fabien Cardinaux, Ghouthi Boukli Hacene, Javier Alonso Garcia