Strong Generalization
Strong generalization, the ability of machine learning models to perform well on unseen data, is a central objective in current research. Active areas of investigation include improving the robustness of self-supervised learning, understanding the optimization dynamics of transformers and other architectures (including CNNs and RNNs), and developing methods to enhance generalization through data augmentation, regularization techniques (e.g., logical regularization, consistency regularization), and improved training strategies (e.g., few-shot learning, meta-learning). These advancements are crucial for building reliable and adaptable AI systems across diverse applications, from image classification and natural language processing to healthcare and robotics.
Papers
Enhancing Sharpness-Aware Optimization Through Variance Suppression
Bingcong Li, Georgios B. Giannakis
LivDet2023 -- Fingerprint Liveness Detection Competition: Advancing Generalization
Marco Micheletto, Roberto Casula, Giulia Orrù, Simone Carta, Sara Concas, Simone Maurizio La Cava, Julian Fierrez, Gian Luca Marcialis
Improving Generalization in Game Agents with Data Augmentation in Imitation Learning
Derek Yadgaroff, Alessandro Sestini, Konrad Tollmar, Ayca Ozcelikkale, Linus Gisslén
Invariant Learning via Probability of Sufficient and Necessary Causes
Mengyue Yang, Zhen Fang, Yonggang Zhang, Yali Du, Furui Liu, Jean-Francois Ton, Jianhong Wang, Jun Wang
Analysis of the Memorization and Generalization Capabilities of AI Agents: Are Continual Learners Robust?
Minsu Kim, Walid Saad
A Discussion on Generalization in Next-Activity Prediction
Luka Abb, Peter Pfeiffer, Peter Fettke, Jana-Rebecca Rehse
Spoofing attack augmentation: can differently-trained attack models improve generalisation?
Wanying Ge, Xin Wang, Junichi Yamagishi, Massimiliano Todisco, Nicholas Evans
Investigating Zero- and Few-shot Generalization in Fact Verification
Liangming Pan, Yunxiang Zhang, Min-Yen Kan