Strong Generalization
Strong generalization, the ability of machine learning models to perform well on unseen data, is a central objective in current research. Active areas of investigation include improving the robustness of self-supervised learning, understanding the optimization dynamics of transformers and other architectures (including CNNs and RNNs), and developing methods to enhance generalization through data augmentation, regularization techniques (e.g., logical regularization, consistency regularization), and improved training strategies (e.g., few-shot learning, meta-learning). These advancements are crucial for building reliable and adaptable AI systems across diverse applications, from image classification and natural language processing to healthcare and robotics.
Papers
Precision, Stability, and Generalization: A Comprehensive Assessment of RNNs learnability capability for Classifying Counter and Dyck Languages
Neisarg Dave, Daniel Kifer, Lee Giles, Ankur Mali
The Effect of Acute Stress on the Interpretability and Generalization of Schizophrenia Predictive Machine Learning Models
Gideon Vos, Maryam Ebrahimpour, Liza van Eijk, Zoltan Sarnyai, Mostafa Rahimi Azghadi
Geometric Collaborative Filtering with Convergence
Hisham Husain, Julien Monteil
OOD-Chameleon: Is Algorithm Selection for OOD Generalization Learnable?
Liangze Jiang, Damien Teney
AlphaIntegrator: Transformer Action Search for Symbolic Integration Proofs
Mert Ünsal, Timon Gehr, Martin Vechev
Generalization emerges from local optimization in a self-organized learning network
S. Barland, L. Gil
Boosting Sample Efficiency and Generalization in Multi-agent Reinforcement Learning via Equivariance
Joshua McClellan, Naveed Haghani, John Winder, Furong Huang, Pratap Tokekar
Achieving Generalization in Orchestrating GNSS Interference Monitoring Stations Through Pseudo-Labeling
Lucas Heublein, Tobias Feigl, Alexander Rügamer, Felix Ott
Simplicity bias and optimization threshold in two-layer ReLU networks
Etienne Boursier, Nicolas Flammarion
Theoretical Insights into Fine-Tuning Attention Mechanism: Generalization and Optimization
Xinhao Yao, Hongjin Qian, Xiaolin Hu, Gengze Xu, Yong Liu
Can Models Learn Skill Composition from Examples?
Haoyu Zhao, Simran Kaur, Dingli Yu, Anirudh Goyal, Sanjeev Arora
Focus On What Matters: Separated Models For Visual-Based RL Generalization
Di Zhang, Bowen Lv, Hai Zhang, Feifan Yang, Junqiao Zhao, Hang Yu, Chang Huang, Hongtu Zhou, Chen Ye, Changjun Jiang