Strong Generalization
Strong generalization, the ability of machine learning models to perform well on unseen data, is a central objective in current research. Active areas of investigation include improving the robustness of self-supervised learning, understanding the optimization dynamics of transformers and other architectures (including CNNs and RNNs), and developing methods to enhance generalization through data augmentation, regularization techniques (e.g., logical regularization, consistency regularization), and improved training strategies (e.g., few-shot learning, meta-learning). These advancements are crucial for building reliable and adaptable AI systems across diverse applications, from image classification and natural language processing to healthcare and robotics.
Papers
Generalization Across Experimental Parameters in Machine Learning Analysis of High Resolution Transmission Electron Microscopy Datasets
Katherine Sytwu, Luis Rangel DaCosta, Mary C. Scott
Blackbird language matrices (BLM), a new task for rule-like generalization in neural networks: Motivations and Formal Specifications
Paola Merlo
Feed Two Birds with One Scone: Exploiting Wild Data for Both Out-of-Distribution Generalization and Detection
Haoyue Bai, Gregory Canal, Xuefeng Du, Jeongyeol Kwon, Robert Nowak, Yixuan Li
Language Aligned Visual Representations Predict Human Behavior in Naturalistic Learning Tasks
Can Demircan, Tankred Saanum, Leonardo Pettini, Marcel Binz, Blazej M Baczkowski, Paula Kaanders, Christian F Doeller, Mona M Garvert, Eric Schulz
PeFLL: Personalized Federated Learning by Learning to Learn
Jonathan Scott, Hossein Zakerinia, Christoph H. Lampert
On the Importance of Exploration for Generalization in Reinforcement Learning
Yiding Jiang, J. Zico Kolter, Roberta Raileanu
In-Context Learning through the Bayesian Prism
Madhur Panwar, Kabir Ahuja, Navin Goyal