Adversarial Example
Adversarial examples are subtly altered inputs designed to fool machine learning models, primarily deep neural networks (DNNs), into making incorrect predictions. Current research focuses on improving model robustness against these attacks, exploring techniques like ensemble methods, multi-objective representation learning, and adversarial training, often applied to architectures such as ResNets and Vision Transformers. Understanding and mitigating the threat of adversarial examples is crucial for ensuring the reliability and security of AI systems across diverse applications, from image classification and natural language processing to malware detection and autonomous driving. The development of robust defenses and effective attack detection methods remains a significant area of ongoing investigation.
Papers
R-LPIPS: An Adversarially Robust Perceptual Similarity Metric
Sara Ghazanfari, Siddharth Garg, Prashanth Krishnamurthy, Farshad Khorrami, Alexandre Araujo
NSA: Naturalistic Support Artifact to Boost Network Confidence
Abhijith Sharma, Phil Munz, Apurva Narayan
FLARE: Fingerprinting Deep Reinforcement Learning Agents using Universal Adversarial Masks
Buse G. A. Tekgul, N. Asokan
Set-level Guidance Attack: Boosting Adversarial Transferability of Vision-Language Pre-training Models
Dong Lu, Zhiqiang Wang, Teng Wang, Weili Guan, Hongchang Gao, Feng Zheng
Enhanced Security against Adversarial Examples Using a Random Ensemble of Encrypted Vision Transformer Models
Ryota Iijima, Miki Tanaka, Sayaka Shiota, Hitoshi Kiya
Lost In Translation: Generating Adversarial Examples Robust to Round-Trip Translation
Neel Bhandari, Pin-Yu Chen
Gradient-Based Word Substitution for Obstinate Adversarial Examples Generation in Language Models
Yimu Wang, Peng Shi, Hongyang Zhang
AdvDiff: Generating Unrestricted Adversarial Examples using Diffusion Models
Xuelong Dai, Kaisheng Liang, Bin Xiao
Robust Automatic Speech Recognition via WavAugment Guided Phoneme Adversarial Training
Gege Qi, Yuefeng Chen, Xiaofeng Mao, Xiaojun Jia, Ranjie Duan, Rong Zhang, Hui Xue
Unveiling Vulnerabilities in Interpretable Deep Learning Systems with Query-Efficient Black-box Attacks
Eldor Abdukhamidov, Mohammed Abuhamad, Simon S. Woo, Eric Chan-Tin, Tamer Abuhmed
Robust Feature Inference: A Test-time Defense Strategy using Spectral Projections
Anurag Singh, Mahalakshmi Sabanayagam, Krikamol Muandet, Debarghya Ghoshdastidar
Improving Transferability of Adversarial Examples via Bayesian Attacks
Qizhang Li, Yiwen Guo, Xiaochen Yang, Wangmeng Zuo, Hao Chen