Explanation Shift

Explanation shift focuses on understanding how changes in input data distributions affect not only the predictive performance of machine learning models but also the characteristics of their explanations. Current research investigates how shifts in explanations can better indicate out-of-distribution behavior than traditional performance metrics, exploring techniques like modeling explanation characteristics and leveraging group-aware optimization to improve explanation robustness. This research is significant because it enhances the reliability and interpretability of machine learning models, particularly in dynamic environments where data distributions are prone to change, leading to more trustworthy and robust AI systems.

Papers