Native Robustness
Native robustness in machine learning focuses on developing models inherently resistant to various forms of input perturbations, including adversarial attacks and noisy data, without relying solely on post-hoc defenses. Current research emphasizes techniques like ensemble methods, reprogramming existing models, and modifying training procedures (e.g., using different learning rates for specific model layers or incorporating regularization methods) to improve robustness across diverse model architectures, including convolutional neural networks, vision transformers, and large language models. This field is crucial for deploying reliable AI systems in safety-critical applications, such as healthcare and autonomous driving, where model resilience to unexpected inputs is paramount.
Papers
Exploring Loss Design Techniques For Decision Tree Robustness To Label Noise
Lukasz Sztukiewicz, Jack Henry Good, Artur Dubrawski
A One-Layer Decoder-Only Transformer is a Two-Layer RNN: With an Application to Certified Robustness
Yuhao Zhang, Aws Albarghouthi, Loris D'Antoni
Improving Data-aware and Parameter-aware Robustness for Continual Learning
Hanxi Xiao, Fan Lyu
Verifying Properties of Binary Neural Networks Using Sparse Polynomial Optimization
Jianting Yang, Srećko Ðurašinović, Jean-Bernard Lasserre, Victor Magron, Jun Zhao
Is Algorithmic Stability Testable? A Unified Framework under Computational Constraints
Yuetian Luo, Rina Foygel Barber
Leveraging Real Electric Guitar Tones and Effects to Improve Robustness in Guitar Tablature Transcription Modeling
Hegel Pedroza, Wallace Abreu, Ryan Corey, Iran Roman
SLIFER: Investigating Performance and Robustness of Malware Detection Pipelines
Andrea Ponte, Dmitrijs Trizna, Luca Demetrio, Battista Biggio, Ivan Tesfai Ogbu, Fabio Roli
The Vital Role of Gradient Clipping in Byzantine-Resilient Distributed Learning
Youssef Allouah, Rachid Guerraoui, Nirupam Gupta, Ahmed Jellouli, Geovani Rizk, John Stephan
Certified Robustness against Sparse Adversarial Perturbations via Data Localization
Ambar Pal, René Vidal, Jeremias Sulam
On the stability of gradient descent with second order dynamics for time-varying cost functions
Travis E. Gibson, Sawal Acharya, Anjali Parashar, Joseph E. Gaudio, Anurdha M. Annaswamy
WaterPool: A Watermark Mitigating Trade-offs among Imperceptibility, Efficacy and Robustness
Baizhou Huang, Xiaojun Wan