Adversarial Example
Adversarial examples are subtly altered inputs designed to fool machine learning models, primarily deep neural networks (DNNs), into making incorrect predictions. Current research focuses on improving model robustness against these attacks, exploring techniques like ensemble methods, multi-objective representation learning, and adversarial training, often applied to architectures such as ResNets and Vision Transformers. Understanding and mitigating the threat of adversarial examples is crucial for ensuring the reliability and security of AI systems across diverse applications, from image classification and natural language processing to malware detection and autonomous driving. The development of robust defenses and effective attack detection methods remains a significant area of ongoing investigation.
Papers
Adversarial Examples on Segmentation Models Can be Easy to Transfer
Jindong Gu, Hengshuang Zhao, Volker Tresp, Philip Torr
Evaluating Adversarial Attacks on ImageNet: A Reality Check on Misclassification Classes
Utku Ozbulak, Maura Pintor, Arnout Van Messem, Wesley De Neve
Medical Aegis: Robust adversarial protectors for medical images
Qingsong Yao, Zecheng He, S. Kevin Zhou
Meta Adversarial Perturbations
Chia-Hung Yuan, Pin-Yu Chen, Chia-Mu Yu
Enhanced countering adversarial attacks via input denoising and feature restoring
Yanni Li, Wenhui Zhang, Jiawei Liu, Xiaoli Kou, Hui Li, Jiangtao Cui
Towards Efficiently Evaluating the Robustness of Deep Neural Networks in IoT Systems: A GAN-based Method
Tao Bai, Jun Zhao, Jinlin Zhu, Shoudong Han, Jiefeng Chen, Bo Li, Alex Kot
TnT Attacks! Universal Naturalistic Adversarial Patches Against Deep Neural Network Systems
Bao Gia Doan, Minhui Xue, Shiqing Ma, Ehsan Abbasnejad, Damith C. Ranasinghe