Adversarial Example
Adversarial examples are subtly altered inputs designed to fool machine learning models, primarily deep neural networks (DNNs), into making incorrect predictions. Current research focuses on improving model robustness against these attacks, exploring techniques like ensemble methods, multi-objective representation learning, and adversarial training, often applied to architectures such as ResNets and Vision Transformers. Understanding and mitigating the threat of adversarial examples is crucial for ensuring the reliability and security of AI systems across diverse applications, from image classification and natural language processing to malware detection and autonomous driving. The development of robust defenses and effective attack detection methods remains a significant area of ongoing investigation.
Papers
Arabic Synonym BERT-based Adversarial Examples for Text Classification
Norah Alshahrani, Saied Alshahrani, Esma Wali, Jeanna Matthews
Transcending Adversarial Perturbations: Manifold-Aided Adversarial Examples with Legitimate Semantics
Shuai Li, Xiaoyu Jiang, Xiaoguang Ma
A Generative Approach to Surrogate-based Black-box Attacks
Raha Moraffah, Huan Liu
$σ$-zero: Gradient-based Optimization of $\ell_0$-norm Adversarial Examples
Antonio Emanuele Cinà, Francesco Villani, Maura Pintor, Lea Schönherr, Battista Biggio, Marcello Pelillo
HQA-Attack: Toward High Quality Black-Box Hard-Label Adversarial Attack on Text
Han Liu, Zhi Xu, Xiaotong Zhang, Feng Zhang, Fenglong Ma, Hongyang Chen, Hong Yu, Xianchao Zhang
STAA-Net: A Sparse and Transferable Adversarial Attack for Speech Emotion Recognition
Yi Chang, Zhao Ren, Zixing Zhang, Xin Jing, Kun Qian, Xi Shao, Bin Hu, Tanja Schultz, Björn W. Schuller
How Robust Are Energy-Based Models Trained With Equilibrium Propagation?
Siddharth Mansingh, Michal Kucer, Garrett Kenyon, Juston Moore, Michael Teti
Finding a Needle in the Adversarial Haystack: A Targeted Paraphrasing Approach For Uncovering Edge Cases with Minimal Distribution Distortion
Aly M. Kassem, Sherif Saad
Robust Localization of Key Fob Using Channel Impulse Response of Ultra Wide Band Sensors for Keyless Entry Systems
Abhiram Kolli, Filippo Casamassima, Horst Possegger, Horst Bischof
Bag of Tricks to Boost Adversarial Transferability
Zeliang Zhang, Rongyi Zhu, Wei Yao, Xiaosen Wang, Chenliang Xu
A Generative Adversarial Attack for Multilingual Text Classifiers
Tom Roth, Inigo Jauregi Unanue, Alsharif Abuadbba, Massimo Piccardi