Native Robustness
Native robustness in machine learning focuses on developing models inherently resistant to various forms of input perturbations, including adversarial attacks and noisy data, without relying solely on post-hoc defenses. Current research emphasizes techniques like ensemble methods, reprogramming existing models, and modifying training procedures (e.g., using different learning rates for specific model layers or incorporating regularization methods) to improve robustness across diverse model architectures, including convolutional neural networks, vision transformers, and large language models. This field is crucial for deploying reliable AI systems in safety-critical applications, such as healthcare and autonomous driving, where model resilience to unexpected inputs is paramount.
Papers
Coalitions of Large Language Models Increase the Robustness of AI Agents
Prattyush Mangal, Carol Mak, Theo Kanakis, Timothy Donovan, Dave Braines, Edward Pyzer-Knapp
Assessing Robustness of Machine Learning Models using Covariate Perturbations
Arun Prakash R, Anwesha Bhattacharyya, Joel Vaughan, Vijayan N. Nair
Certifiably Robust Encoding Schemes
Aman Saxena, Tom Wollschläger, Nicola Franco, Jeanette Miriam Lorenz, Stephan Günnemann
Adversarial Robust Decision Transformer: Enhancing Robustness of RvS via Minimax Returns-to-go
Xiaohang Tang, Afonso Marques, Parameswaran Kamalaruban, Ilija Bogunovic
Sparse vs Contiguous Adversarial Pixel Perturbations in Multimodal Models: An Empirical Analysis
Cristian-Alexandru Botocan, Raphael Meier, Ljiljana Dolamic
Robustness of Speech Separation Models for Similar-pitch Speakers
Bunlong Lay, Sebastian Zaczek, Kristina Tesch, Timo Gerkmann
Targeted Latent Adversarial Training Improves Robustness to Persistent Harmful Behaviors in LLMs
Abhay Sheshadri, Aidan Ewart, Phillip Guo, Aengus Lynch, Cindy Wu, Vivek Hebbar, Henry Sleight, Asa Cooper Stickland, Ethan Perez, Dylan Hadfield-Menell, Stephen Casper
Increasing the Robustness of Model Predictions to Missing Sensors in Earth Observation
Francisco Mena, Diego Arenas, Andreas Dengel
Craft: Cross-modal Aligned Features Improve Robustness of Prompt Tuning
Jingchen Sun, Rohan Sharma, Vishnu Suresh Lokhande, Changyou Chen
Out of spuriousity: Improving robustness to spurious correlations without group annotations
Phuong Quynh Le, Jörg Schlötterer, Christin Seifert
ARoFace: Alignment Robustness to Improve Low-Quality Face Recognition
Mohammad Saeed Ebrahimi Saadabadi, Sahar Rahimi Malakshan, Ali Dabouei, Nasser M. Nasrabadi