Adversarial Example
Adversarial examples are subtly altered inputs designed to fool machine learning models, primarily deep neural networks (DNNs), into making incorrect predictions. Current research focuses on improving model robustness against these attacks, exploring techniques like ensemble methods, multi-objective representation learning, and adversarial training, often applied to architectures such as ResNets and Vision Transformers. Understanding and mitigating the threat of adversarial examples is crucial for ensuring the reliability and security of AI systems across diverse applications, from image classification and natural language processing to malware detection and autonomous driving. The development of robust defenses and effective attack detection methods remains a significant area of ongoing investigation.
Papers
$DA^3$: A Distribution-Aware Adversarial Attack against Language Models
Yibo Wang, Xiangjue Dong, James Caverlee, Philip S. Yu
Physical Adversarial Examples for Multi-Camera Systems
Ana Răduţoiu, Jan-Philipp Schulze, Philip Sperl, Konstantin Böttinger
On The Relationship Between Universal Adversarial Attacks And Sparse Representations
Dana Weitzner, Raja Giryes
Distilling Out-of-Distribution Robustness from Vision-Language Foundation Models
Andy Zhou, Jindong Wang, Yu-Xiong Wang, Haohan Wang
Upper and lower bounds for the Lipschitz constant of random neural networks
Paul Geuchen, Thomas Heindl, Dominik Stöger, Felix Voigtlaender
Towards Evaluating Transfer-based Attacks Systematically, Practically, and Fairly
Qizhang Li, Yiwen Guo, Wangmeng Zuo, Hao Chen