Inductive Bias
Inductive bias refers to the assumptions built into machine learning models that guide their learning process, influencing what types of solutions they find and how well they generalize to unseen data. Current research focuses on understanding and controlling inductive biases in various model architectures, including neural networks (particularly transformers and graph neural networks), and exploring how biases affect model performance, fairness, and robustness across diverse tasks such as image classification, natural language processing, and reinforcement learning. This research is crucial for improving model generalization, mitigating biases, and developing more efficient and reliable machine learning systems across numerous scientific and practical applications.
Papers
Equivariant Reinforcement Learning under Partial Observability
Hai Nguyen, Andrea Baisero, David Klee, Dian Wang, Robert Platt, Christopher Amato
Rethinking Knowledge Transfer in Learning Using Privileged Information
Danil Provodin, Bram van den Akker, Christina Katsimerou, Maurits Kaptein, Mykola Pechenizkiy