Strong Generalization
Strong generalization, the ability of machine learning models to perform well on unseen data, is a central objective in current research. Active areas of investigation include improving the robustness of self-supervised learning, understanding the optimization dynamics of transformers and other architectures (including CNNs and RNNs), and developing methods to enhance generalization through data augmentation, regularization techniques (e.g., logical regularization, consistency regularization), and improved training strategies (e.g., few-shot learning, meta-learning). These advancements are crucial for building reliable and adaptable AI systems across diverse applications, from image classification and natural language processing to healthcare and robotics.
Papers
Deepfake Sentry: Harnessing Ensemble Intelligence for Resilient Detection and Generalisation
Liviu-Daniel Ştefan, Dan-Cristian Stanciu, Mihai Dogariu, Mihai Gabriel Constantin, Andrei Cosmin Jitaru, Bogdan Ionescu
Development of Compositionality and Generalization through Interactive Learning of Language and Action of Robots
Prasanna Vijayaraghavan, Jeffrey Frederic Queisser, Sergio Verduzco Flores, Jun Tani
Improving Generalization via Meta-Learning on Hard Samples
Nishant Jain, Arun S. Suggala, Pradeep Shenoy
Towards Understanding the Relationship between In-context Learning and Compositional Generalization
Sungjun Han, Sebastian Padó
Towards Generalizing to Unseen Domains with Few Labels
Chamuditha Jayanga Galappaththige, Sanoojan Baliah, Malitha Gunawardhana, Muhammad Haris Khan
Generalization of Scaled Deep ResNets in the Mean-Field Regime
Yihang Chen, Fanghui Liu, Yiping Lu, Grigorios G. Chrysos, Volkan Cevher
Laying the Foundation First? Investigating the Generalization from Atomic Skills to Complex Reasoning Tasks
Yuncheng Huang, Qianyu He, Yipei Xu, Jiaqing Liang, Yanghua Xiao
Agile gesture recognition for low-power applications: customisation for generalisation
Ying Liu, Liucheng Guo, Valeri A. Makarovc, Alexander Gorbana, Evgeny Mirkesa, Ivan Y. Tyukin
Do Agents Dream of Electric Sheep?: Improving Generalization in Reinforcement Learning through Generative Learning
Giorgio Franceschelli, Mirco Musolesi
Input Data Adaptive Learning (IDAL) for Sub-acute Ischemic Stroke Lesion Segmentation
Michael Götz, Christian Weber, Christoph Kolb, Klaus Maier-Hein