Strong Generalization
Strong generalization, the ability of machine learning models to perform well on unseen data, is a central objective in current research. Active areas of investigation include improving the robustness of self-supervised learning, understanding the optimization dynamics of transformers and other architectures (including CNNs and RNNs), and developing methods to enhance generalization through data augmentation, regularization techniques (e.g., logical regularization, consistency regularization), and improved training strategies (e.g., few-shot learning, meta-learning). These advancements are crucial for building reliable and adaptable AI systems across diverse applications, from image classification and natural language processing to healthcare and robotics.
Papers
The ConceptARC Benchmark: Evaluating Understanding and Generalization in the ARC Domain
Arseny Moskvichev, Victor Vikram Odouard, Melanie Mitchell
Generalization bounds for neural ordinary differential equations and deep residual networks
Pierre Marion
Semantic uncertainty guides the extension of conventions to new referents
Ron Eliav, Anya Ji, Yoav Artzi, Robert D. Hawkins
CROP: Towards Distributional-Shift Robust Reinforcement Learning using Compact Reshaped Observation Processing
Philipp Altmann, Fabian Ritz, Leonard Feuchtinger, Jonas Nüßlein, Claudia Linnhoff-Popien, Thomy Phan
Can Agents Run Relay Race with Strangers? Generalization of RL to Out-of-Distribution Trajectories
Li-Cheng Lan, Huan Zhang, Cho-Jui Hsieh