Native Robustness
Native robustness in machine learning focuses on developing models inherently resistant to various forms of input perturbations, including adversarial attacks and noisy data, without relying solely on post-hoc defenses. Current research emphasizes techniques like ensemble methods, reprogramming existing models, and modifying training procedures (e.g., using different learning rates for specific model layers or incorporating regularization methods) to improve robustness across diverse model architectures, including convolutional neural networks, vision transformers, and large language models. This field is crucial for deploying reliable AI systems in safety-critical applications, such as healthcare and autonomous driving, where model resilience to unexpected inputs is paramount.
Papers
Knowledge-Augmented Reasoning for EUAIA Compliance and Adversarial Robustness of LLMs
Tomas Bueno Momcilovic, Dian Balta, Beat Buesser, Giulio Zizzo, Mark Purcell
Make Interval Bound Propagation great again
Patryk Krukowski, Daniel Wilczak, Jacek Tabor, Anna Bielawska, Przemysław Spurek
FedCert: Federated Accuracy Certification
Minh Hieu Nguyen, Huu Tien Nguyen, Trung Thanh Nguyen, Manh Duong Nguyen, Trong Nghia Hoang, Truong Thao Nguyen, Phi Le Nguyen
Learning-Augmented Robust Algorithmic Recourse
Kshitij Kayastha, Vasilis Gkatzelis, Shahin Jabbari
Toward a Holistic Evaluation of Robustness in CLIP Models
Weijie Tu, Weijian Deng, Tom Gedeon
On the Robustness of Machine Learning Models in Predicting Thermodynamic Properties: a Case of Searching for New Quasicrystal Approximants
Fedor S. Avilov, Roman A. Eremin, Semen A. Budennyy, Innokentiy S. Humonen
Efficient PAC Learning of Halfspaces with Constant Malicious Noise Rate
Jie Shen, Xiaoyu Li
EndoDepth: A Benchmark for Assessing Robustness in Endoscopic Depth Prediction
Ivan Reyes-Amezcua, Ricardo Espinosa, Christian Daul, Gilberto Ochoa-Ruiz, Andres Mendez-Vazquez
SATA: Spatial Autocorrelation Token Analysis for Enhancing the Robustness of Vision Transformers
Nick Nikzad, Yi Liao, Yongsheng Gao, Jun Zhou
Enhancing Robustness of Graph Neural Networks through p-Laplacian
Anuj Kumar Sirohi, Subhanu Halder, Kabir Kumar, Sandeep Kumar
Efficient Noise Mitigation for Enhancing Inference Accuracy in DNNs on Mixed-Signal Accelerators
Seyedarmin Azizi, Mohammad Erfan Sadeghi, Mehdi Kamal, Massoud Pedram
Robustness of AI-based weather forecasts in a changing climate
Thomas Rackow, Nikolay Koldunov, Christian Lessig, Irina Sandu, Mihai Alexe, Matthew Chantry, Mariana Clare, Jesper Dramsch, Florian Pappenberger, Xabier Pedruzo-Bagazgoitia, Steffen Tietsche, Thomas Jung
Evaluating the Performance and Robustness of LLMs in Materials Science Q&A and Property Predictions
Hongchen Wang, Kangming Li, Scott Ramsay, Yao Fehlis, Edward Kim, Jason Hattrick-Simpers
ESPERANTO: Evaluating Synthesized Phrases to Enhance Robustness in AI Detection for Text Origination
Navid Ayoobi, Lily Knab, Wen Cheng, David Pantoja, Hamidreza Alikhani, Sylvain Flamant, Jin Kim, Arjun Mukherjee