Adversarial Example
Adversarial examples are subtly altered inputs designed to fool machine learning models, primarily deep neural networks (DNNs), into making incorrect predictions. Current research focuses on improving model robustness against these attacks, exploring techniques like ensemble methods, multi-objective representation learning, and adversarial training, often applied to architectures such as ResNets and Vision Transformers. Understanding and mitigating the threat of adversarial examples is crucial for ensuring the reliability and security of AI systems across diverse applications, from image classification and natural language processing to malware detection and autonomous driving. The development of robust defenses and effective attack detection methods remains a significant area of ongoing investigation.
Papers
RobustFair: Adversarial Evaluation through Fairness Confusion Directed Gradient Search
Xuran Li, Peng Wu, Kaixiang Dong, Zhen Zhang, Yanting Chen
How Deep Learning Sees the World: A Survey on Adversarial Attacks & Defenses
Joana C. Costa, Tiago Roxo, Hugo Proença, Pedro R. M. Inácio
Towards an Accurate and Secure Detector against Adversarial Perturbations
Chao Wang, Shuren Qi, Zhiqiu Huang, Yushu Zhang, Rushi Lan, Xiaochun Cao
Content-based Unrestricted Adversarial Attack
Zhaoyu Chen, Bo Li, Shuang Wu, Kaixun Jiang, Shouhong Ding, Wenqiang Zhang
Toward Adversarial Training on Contextualized Language Representation
Hongqiu Wu, Yongxiang Liu, Hanwen Shi, Hai Zhao, Min Zhang
Adversarial Examples Detection with Enhanced Image Difference Features based on Local Histogram Equalization
Zhaoxia Yin, Shaowei Zhu, Hang Su, Jianteng Peng, Wanli Lyu, Bin Luo