Inductive Bias
Inductive bias refers to the assumptions built into machine learning models that guide their learning process, influencing what types of solutions they find and how well they generalize to unseen data. Current research focuses on understanding and controlling inductive biases in various model architectures, including neural networks (particularly transformers and graph neural networks), and exploring how biases affect model performance, fairness, and robustness across diverse tasks such as image classification, natural language processing, and reinforcement learning. This research is crucial for improving model generalization, mitigating biases, and developing more efficient and reliable machine learning systems across numerous scientific and practical applications.
Papers
Fast Rates for Noisy Interpolation Require Rethinking the Effects of Inductive Bias
Konstantin Donhauser, Nicolo Ruggeri, Stefan Stojanovic, Fanny Yang
Discovering Inductive Bias with Gibbs Priors: A Diagnostic Tool for Approximate Bayesian Inference
Luca Rendsburg, Agustinus Kristiadi, Philipp Hennig, Ulrike von Luxburg