Native Robustness
Native robustness in machine learning focuses on developing models inherently resistant to various forms of input perturbations, including adversarial attacks and noisy data, without relying solely on post-hoc defenses. Current research emphasizes techniques like ensemble methods, reprogramming existing models, and modifying training procedures (e.g., using different learning rates for specific model layers or incorporating regularization methods) to improve robustness across diverse model architectures, including convolutional neural networks, vision transformers, and large language models. This field is crucial for deploying reliable AI systems in safety-critical applications, such as healthcare and autonomous driving, where model resilience to unexpected inputs is paramount.
Papers
Statistical, Robustness, and Computational Guarantees for Sliced Wasserstein Distances
Sloan Nietert, Ritwik Sadhu, Ziv Goldfeld, Kengo Kato
DE-CROP: Data-efficient Certified Robustness for Pretrained Classifiers
Gaurav Kumar Nayak, Ruchit Rawal, Anirban Chakraborty
A.I. Robustness: a Human-Centered Perspective on Technological Challenges and Opportunities
Andrea Tocchetti, Lorenzo Corti, Agathe Balayn, Mireia Yurrita, Philip Lippmann, Marco Brambilla, Jie Yang
Is your noise correction noisy? PLS: Robustness to label noise with two stage detection
Paul Albert, Eric Arazo, Tarun Krishna, Noel E. O'Connor, Kevin McGuinness
Everything is Varied: The Surprising Impact of Individual Variation on ML Robustness in Medicine
Andrea Campagner, Lorenzo Famiglini, Anna Carobene, Federico Cabitza
Robustness of Unsupervised Representation Learning without Labels
Aleksandar Petrov, Marta Kwiatkowska
Enhance Sample Efficiency and Robustness of End-to-end Urban Autonomous Driving via Semantic Masked World Model
Zeyu Gao, Yao Mu, Chen Chen, Jingliang Duan, Shengbo Eben Li, Ping Luo, Yanfeng Lu
ViewFool: Evaluating the Robustness of Visual Recognition to Adversarial Viewpoints
Yinpeng Dong, Shouwei Ruan, Hang Su, Caixin Kang, Xingxing Wei, Jun Zhu
1st ICLR International Workshop on Privacy, Accountability, Interpretability, Robustness, Reasoning on Structured Data (PAIR^2Struct)
Hao Wang, Wanyu Lin, Hao He, Di Wang, Chengzhi Mao, Muhan Zhang
A2: Efficient Automated Attacker for Boosting Adversarial Training
Zhuoer Xu, Guanghui Zhu, Changhua Meng, Shiwen Cui, Zhenzhe Ying, Weiqiang Wang, Ming GU, Yihua Huang