Robustness Evaluation Framework

Robustness evaluation frameworks aim to assess the reliability and stability of machine learning models, particularly deep learning models, under various perturbations or adversarial attacks. Current research focuses on developing standardized evaluation methods across different modalities (image, text, etc.), encompassing diverse attack types and incorporating metrics that capture both accuracy and the stability of model outputs or explanations. These frameworks are crucial for building trustworthy AI systems, improving model generalization, and ensuring reliable performance in real-world applications where noisy or manipulated inputs are common.

Papers