Generalization Bound
Generalization bounds in machine learning aim to quantify a model's ability to perform well on unseen data, based on its performance on training data. Current research focuses on developing tighter bounds for various architectures, including neural networks (especially deep and "nearly-linear" networks), large language models, and graph neural networks, often employing techniques like sample compression, PAC-Bayesian analysis, and information-theoretic approaches. These advancements are crucial for understanding and improving the reliability and robustness of machine learning models, particularly in high-stakes applications where generalization is paramount. The development of practically computable and informative bounds remains a significant challenge and active area of investigation.
Papers
Task-Free Continual Learning via Online Discrepancy Distance Learning
Fei Ye, Adrian G. Bors
Evaluated CMI Bounds for Meta Learning: Tightness and Expressiveness
Fredrik Hellström, Giuseppe Durisi
A New Family of Generalization Bounds Using Samplewise Evaluated CMI
Fredrik Hellström, Giuseppe Durisi
On the Importance of Gradient Norm in PAC-Bayesian Bounds
Itai Gat, Yossi Adi, Alexander Schwing, Tamir Hazan