Adversarial Sample
Adversarial samples are inputs designed to intentionally mislead machine learning models, primarily by introducing small, imperceptible perturbations to otherwise correctly classified data. Current research focuses on developing more robust models through techniques like adversarial training, purification methods using generative models (e.g., GANs), and exploring the vulnerabilities of various architectures, including convolutional neural networks, recurrent networks, and large language models. Understanding and mitigating the impact of adversarial samples is crucial for ensuring the reliability and security of machine learning systems across diverse applications, from cybersecurity to medical diagnosis.
Papers
Transferable Adversarial Attacks on Vision Transformers with Token Gradient Regularization
Jianping Zhang, Yizhan Huang, Weibin Wu, Michael R. Lyu
Improving the Transferability of Adversarial Samples by Path-Augmented Method
Jianping Zhang, Jen-tse Huang, Wenxuan Wang, Yichen Li, Weibin Wu, Xiaosen Wang, Yuxin Su, Michael R. Lyu
Test-time Detection and Repair of Adversarial Samples via Masked Autoencoder
Yun-Yun Tsai, Ju-Chin Chao, Albert Wen, Zhaoyuan Yang, Chengzhi Mao, Tapan Shah, Junfeng Yang
Wasserstein Loss for Semantic Editing in the Latent Space of GANs
Perla Doubinsky, Nicolas Audebert, Michel Crucianu, Hervé Le Borgne