Robustness Evaluation
Robustness evaluation assesses the reliability and stability of machine learning models under various perturbations and unexpected inputs, aiming to ensure their safe and effective deployment in real-world applications. Current research focuses on developing comprehensive benchmarks and metrics to evaluate robustness across diverse domains, including natural language processing, computer vision, and reinforcement learning, often employing adversarial attacks and data augmentation techniques to stress-test models. This field is crucial for building trustworthy AI systems, as robust models are less susceptible to errors and failures caused by noisy data, adversarial attacks, or unexpected environmental conditions, ultimately improving the safety and reliability of AI-driven technologies.
Papers
StaDRe and StaDRo: Reliability and Robustness Estimation of ML-based Forecasting using Statistical Distance Measures
Mohammed Naveed Akram, Akshatha Ambekar, Ioannis Sorokos, Koorosh Aslansefat, Daniel Schneider
Is Multi-Modal Necessarily Better? Robustness Evaluation of Multi-modal Fake News Detection
Jinyin Chen, Chengyu Jia, Haibin Zheng, Ruoxi Chen, Chenbo Fu