Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
Papers
Towards the Transferable Audio Adversarial Attack via Ensemble Methods
Feng Guo, Zheng Sun, Yuxuan Chen, Lei Ju
Masked Language Model Based Textual Adversarial Example Detection
Xiaomei Zhang, Zhaoxi Zhang, Qi Zhong, Xufei Zheng, Yanjun Zhang, Shengshan Hu, Leo Yu Zhang
AI Product Security: A Primer for Developers
Ebenezer R. H. P. Isaac, Jim Reno
Defense-Prefix for Preventing Typographic Attacks on CLIP
Hiroki Azuma, Yusuke Matsui
Generating Adversarial Attacks in the Latent Space
Nitish Shukla, Sudipta Banerjee
Certifiable Black-Box Attacks with Randomized Adversarial Examples: Breaking Defenses with Provable Confidence
Hanbin Hong, Xinyu Zhang, Binghui Wang, Zhongjie Ba, Yuan Hong
RobCaps: Evaluating the Robustness of Capsule Networks against Affine Transformations and Adversarial Attacks
Alberto Marchisio, Antonio De Marco, Alessio Colucci, Maurizio Martina, Muhammad Shafique
Benchmarking the Robustness of Quantized Models
Yisong Xiao, Tianyuan Zhang, Shunchang Liu, Haotong Qin
Robust Deep Learning Models Against Semantic-Preserving Adversarial Attack
Dashan Gao, Yunce Zhao, Yinghua Yao, Zeqi Zhang, Bifei Mao, Xin Yao