Model Sensitivity

Model sensitivity, the degree to which a model's output changes in response to variations in its inputs or parameters, is a crucial area of research across diverse machine learning applications. Current investigations focus on identifying and mitigating sensitivity to spurious correlations, biases (e.g., gender bias in language models), and artifacts in data, employing techniques like spectral analysis, gradient penalization, and surrogate models. Understanding and controlling model sensitivity is vital for improving model robustness, generalization, and reliability, ultimately leading to more trustworthy and effective AI systems in various fields, from medical imaging to climate modeling.

Papers