Native Robustness
Native robustness in machine learning focuses on developing models inherently resistant to various forms of input perturbations, including adversarial attacks and noisy data, without relying solely on post-hoc defenses. Current research emphasizes techniques like ensemble methods, reprogramming existing models, and modifying training procedures (e.g., using different learning rates for specific model layers or incorporating regularization methods) to improve robustness across diverse model architectures, including convolutional neural networks, vision transformers, and large language models. This field is crucial for deploying reliable AI systems in safety-critical applications, such as healthcare and autonomous driving, where model resilience to unexpected inputs is paramount.
Papers
How Does Frequency Bias Affect the Robustness of Neural Image Classifiers against Common Corruption and Adversarial Perturbations?
Alvin Chan, Yew-Soon Ong, Clement Tan
Robustness of Humans and Machines on Object Recognition with Extreme Image Transformations
Dakarai Crowder, Girik Malik
Btech thesis report on adversarial attack detection and purification of adverserially attacked images
Dvij Kalaria
Formulating Robustness Against Unforeseen Attacks
Sihui Dai, Saeed Mahloujifar, Prateek Mittal
Randomized Smoothing under Attack: How Good is it in Pratice?
Thibault Maho, Teddy Furon, Erwan Le Merrer
Improving the Robustness of Federated Learning for Severely Imbalanced Datasets
Debasrita Chakraborty, Ashish Ghosh
Improving robustness of language models from a geometry-aware perspective
Bin Zhu, Zhaoquan Gu, Le Wang, Jinyin Chen, Qi Xuan