Invariant Learning Method
Invariant learning aims to develop machine learning models that are robust to variations in data distribution while preserving predictive accuracy. Current research focuses on improving the efficiency and effectiveness of invariant learning through novel algorithms like frame averaging and co-mixup strategies, often applied to Gaussian processes and graph neural networks, and addressing the challenge of environment partitioning for optimal performance. This field is crucial for enhancing the generalization capabilities of models, particularly in out-of-distribution scenarios, and has significant implications for various applications, including robotics and safety-critical systems.
Papers
Domain Invariant Learning for Gaussian Processes and Bayesian Exploration
Xilong Zhao, Siyuan Bian, Yaoyun Zhang, Yuliang Zhang, Qinying Gu, Xinbing Wang, Chenghu Zhou, Nanyang Ye
Graph Invariant Learning with Subgraph Co-mixup for Out-Of-Distribution Generalization
Tianrui Jia, Haoyang Li, Cheng Yang, Tao Tao, Chuan Shi