Generalization Performance
Generalization performance in machine learning focuses on a model's ability to accurately predict outcomes on unseen data, a crucial aspect for real-world applications. Current research investigates this through various lenses, including mitigating overfitting in self-supervised and federated learning, improving the robustness of models to out-of-distribution data (e.g., using dropout or orthogonal regularization), and enhancing the efficiency of fine-tuning large pre-trained models (e.g., via low-rank adaptation). Understanding and improving generalization is vital for building reliable and adaptable AI systems across diverse domains, impacting fields from image recognition and natural language processing to control of biological neural networks.
Papers
Using Focal Loss to Fight Shallow Heuristics: An Empirical Analysis of Modulated Cross-Entropy in Natural Language Inference
Frano Rajič, Ivan Stresec, Axel Marmet, Tim Poštuvan
Relating Regularization and Generalization through the Intrinsic Dimension of Activations
Bradley C. A. Brown, Jordan Juravsky, Anthony L. Caterini, Gabriel Loaiza-Ganem
GULP: a prediction-based metric between representations
Enric Boix-Adsera, Hannah Lawrence, George Stepaniants, Philippe Rigollet
Guaranteed Conservation of Momentum for Learning Particle-based Fluid Dynamics
Lukas Prantl, Benjamin Ummenhofer, Vladlen Koltun, Nils Thuerey
Optimizing Evaluation Metrics for Multi-Task Learning via the Alternating Direction Method of Multipliers
Ge-Yang Ke, Yan Pan, Jian Yin, Chang-Qin Huang