Native Robustness
Native robustness in machine learning focuses on developing models inherently resistant to various forms of input perturbations, including adversarial attacks and noisy data, without relying solely on post-hoc defenses. Current research emphasizes techniques like ensemble methods, reprogramming existing models, and modifying training procedures (e.g., using different learning rates for specific model layers or incorporating regularization methods) to improve robustness across diverse model architectures, including convolutional neural networks, vision transformers, and large language models. This field is crucial for deploying reliable AI systems in safety-critical applications, such as healthcare and autonomous driving, where model resilience to unexpected inputs is paramount.
Papers
Quantifying the robustness of deep multispectral segmentation models against natural perturbations and data poisoning
Elise Bishoff, Charles Godfrey, Myles McKay, Eleanor Byler
Writing your own book: A method for going from closed to open book QA to improve robustness and performance of smaller LLMs
Giorgi Kokaia, Pratyush Sinha, Yutong Jiang, Nozha Boujemaa
RobustFair: Adversarial Evaluation through Fairness Confusion Directed Gradient Search
Xuran Li, Peng Wu, Kaixiang Dong, Zhen Zhang, Yanting Chen
Raising the Bar for Certified Adversarial Robustness with Diffusion Models
Thomas Altstidl, David Dobre, Björn Eskofier, Gauthier Gidel, Leo Schwinn
rWISDM: Repaired WISDM, a Public Dataset for Human Activity Recognition
Mohammadreza Heydarian, Thomas E. Doyle
On the ISS Property of the Gradient Flow for Single Hidden-Layer Neural Networks with Linear Activations
Arthur Castello B. de Oliveira, Milad Siami, Eduardo D. Sontag
Assessing Hidden Risks of LLMs: An Empirical Study on Robustness, Consistency, and Credibility
Wentao Ye, Mingfeng Ou, Tianyi Li, Yipeng chen, Xuetao Ma, Yifan Yanggong, Sai Wu, Jie Fu, Gang Chen, Haobo Wang, Junbo Zhao
Sensitivity and Robustness of Large Language Models to Prompt Template in Japanese Text Classification Tasks
Chengguang Gan, Tatsunori Mori
An Empirical Study on the Robustness of the Segment Anything Model (SAM)
Yuqing Wang, Yun Zhao, Linda Petzold
A Survey on the Robustness of Computer Vision Models against Common Corruptions
Shunxin Wang, Raymond Veldhuis, Christoph Brune, Nicola Strisciuglio
Causal Information Splitting: Engineering Proxy Features for Robustness to Distribution Shifts
Bijan Mazaheri, Atalanti Mastakouri, Dominik Janzing, Michaela Hardt