Adversarial Example
Adversarial examples are subtly altered inputs designed to fool machine learning models, primarily deep neural networks (DNNs), into making incorrect predictions. Current research focuses on improving model robustness against these attacks, exploring techniques like ensemble methods, multi-objective representation learning, and adversarial training, often applied to architectures such as ResNets and Vision Transformers. Understanding and mitigating the threat of adversarial examples is crucial for ensuring the reliability and security of AI systems across diverse applications, from image classification and natural language processing to malware detection and autonomous driving. The development of robust defenses and effective attack detection methods remains a significant area of ongoing investigation.
Papers
Semi-Implicit Hybrid Gradient Methods with Application to Adversarial Robustness
Beomsu Kim, Junghoon Seo
HoneyModels: Machine Learning Honeypots
Ahmed Abdou, Ryan Sheatsley, Yohan Beugin, Tyler Shipp, Patrick McDaniel
Transferring Adversarial Robustness Through Robust Representation Matching
Pratik Vaishnavi, Kevin Eykholt, Amir Rahmati
Adversarial Attacks and Defense Methods for Power Quality Recognition
Jiwei Tian, Buhong Wang, Jing Li, Zhen Wang, Mete Ozay
Predicting Out-of-Distribution Error with the Projection Norm
Yaodong Yu, Zitong Yang, Alexander Wei, Yi Ma, Jacob Steinhardt
FAAG: Fast Adversarial Audio Generation through Interactive Attack Optimisation
Yuantian Miao, Chao Chen, Lei Pan, Jun Zhang, Yang Xiang
Visualizing Automatic Speech Recognition -- Means for a Better Understanding?
Karla Markert, Romain Parracone, Mykhailo Kulakov, Philip Sperl, Ching-Yu Kao, Konstantin Böttinger
Language Dependencies in Adversarial Attacks on Speech Recognition Systems
Karla Markert, Donika Mirdita, Konstantin Böttinger