Stronger Generalizability
Stronger generalizability in machine learning models is a crucial research area aiming to improve the ability of models trained on one dataset to perform well on unseen data or tasks. Current efforts focus on developing robust methodologies for model evaluation, exploring architectures like Graph Neural Networks and transformers, and investigating techniques such as prompt engineering, data augmentation, and ensemble methods to enhance model performance across diverse scenarios. This pursuit is vital for building reliable and trustworthy AI systems applicable across various domains, from healthcare and drug discovery to robotics and environmental monitoring, ultimately increasing the impact and practical utility of machine learning.
Papers
RealBehavior: A Framework for Faithfully Characterizing Foundation Models' Human-like Behavior Mechanisms
Enyu Zhou, Rui Zheng, Zhiheng Xi, Songyang Gao, Xiaoran Fan, Zichu Fei, Jingting Ye, Tao Gui, Qi Zhang, Xuanjing Huang
Generalizability of CNN Architectures for Face Morph Presentation Attack
Sherko R. HmaSalah, Aras Asaad
Towards Understanding the Generalizability of Delayed Stochastic Gradient Descent
Xiaoge Deng, Li Shen, Shengwei Li, Tao Sun, Dongsheng Li, Dacheng Tao
Open-vocabulary Video Question Answering: A New Benchmark for Evaluating the Generalizability of Video Question Answering Models
Dohwan Ko, Ji Soo Lee, Miso Choi, Jaewon Chu, Jihwan Park, Hyunwoo J. Kim