Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
Papers
Adversarial Prompt Tuning for Vision-Language Models
Jiaming Zhang, Xingjun Ma, Xin Wang, Lingyu Qiu, Jiaqi Wang, Yu-Gang Jiang, Jitao Sang
Robust Network Slicing: Multi-Agent Policies, Adversarial Attacks, and Defensive Strategies
Feng Wang, M. Cenk Gursoy, Senem Velipasalar
Attention-Based Real-Time Defenses for Physical Adversarial Attacks in Vision Applications
Giulio Rossolini, Alessandro Biondi, Giorgio Buttazzo
Hijacking Large Language Models via Adversarial In-Context Learning
Yao Qiang, Xiangyu Zhou, Dongxiao Zhu
Breaking Boundaries: Balancing Performance and Robustness in Deep Wireless Traffic Forecasting
Romain Ilbert, Thai V. Hoang, Zonghua Zhang, Themis Palpanas
Bergeron: Combating Adversarial Attacks through a Conscience-Based Alignment Framework
Matthew Pisano, Peter Ly, Abraham Sanders, Bingsheng Yao, Dakuo Wang, Tomek Strzalkowski, Mei Si
$DA^3$: A Distribution-Aware Adversarial Attack against Language Models
Yibo Wang, Xiangjue Dong, James Caverlee, Philip S. Yu
Alignment is not sufficient to prevent large language models from generating harmful information: A psychoanalytic perspective
Zi Yin, Wei Ding, Jia Liu
The Perception-Robustness Tradeoff in Deterministic Image Restoration
Guy Ohayon, Tomer Michaeli, Michael Elad
On The Relationship Between Universal Adversarial Attacks And Sparse Representations
Dana Weitzner, Raja Giryes
Towards Improving Robustness Against Common Corruptions in Object Detectors Using Adversarial Contrastive Learning
Shashank Kotyan, Danilo Vasconcellos Vargas
An Extensive Study on Adversarial Attack against Pre-trained Models of Code
Xiaohu Du, Ming Wen, Zichao Wei, Shangwen Wang, Hai Jin
Hardest Monotone Functions for Evolutionary Algorithms
Marc Kaufmann, Maxime Larcher, Johannes Lengler, Oliver Sieberling
Multi-agent Attacks for Black-box Social Recommendations
Shijie Wang, Wenqi Fan, Xiao-yong Wei, Xiaowei Mei, Shanru Lin, Qing Li
Transferability Bound Theory: Exploring Relationship between Adversarial Transferability and Flatness
Mingyuan Fan, Xiaodan Li, Cen Chen, Wenmeng Zhou, Yaliang Li
Fight Fire with Fire: Combating Adversarial Patch Attacks using Pattern-randomized Defensive Patches
Jianan Feng, Jiachun Li, Changqing Miao, Jianjun Huang, Wei You, Wenchang Shi, Bin Liang