Sanity Check

Sanity checks in machine learning and related fields aim to verify the reliability and trustworthiness of model explanations and predictions, addressing concerns about potential biases, manipulation, or inaccuracies. Current research focuses on developing and refining these checks for various applications, including image generation detection, saliency map interpretation, and time series classification, often employing techniques like model parameter randomization and data randomization tests, as well as comparing against simpler baseline models. These efforts are crucial for improving the transparency and accountability of AI systems, ultimately leading to more robust and reliable models across diverse scientific and practical domains.

Papers