Model Overfitting
Model overfitting, where a machine learning model learns the training data too well, hindering its ability to generalize to unseen data, remains a central challenge in the field. Current research focuses on understanding and mitigating overfitting in various contexts, including self-supervised learning, large language models, and continual learning, often employing techniques like regularization, ensemble methods, and data augmentation tailored to specific architectures (e.g., transformers, convolutional neural networks). Addressing overfitting is crucial for improving the reliability and robustness of machine learning models across diverse applications, from image recognition and natural language processing to scientific discovery and engineering design. This ongoing work aims to develop more generalizable and trustworthy AI systems.
Papers
Accurate Ground-Truth Depth Image Generation via Overfit Training of Point Cloud Registration using Local Frame Sets
Jiwan Kim, Minchang Kim, Yeong-Gil Shin, Minyoung Chung
Have we been Naive to Select Machine Learning Models? Noisy Data are here to Stay!
Felipe Costa Farias, Teresa Bernarda Ludermir, Carmelo José Albanez Bastos-Filho
Benign, Tempered, or Catastrophic: A Taxonomy of Overfitting
Neil Mallinar, James B. Simon, Amirhesam Abedsoltan, Parthe Pandit, Mikhail Belkin, Preetum Nakkiran