Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
Papers
Fall Leaf Adversarial Attack on Traffic Sign Classification
Anthony Etim, Jakub Szefer
An indicator for effectiveness of text-to-image guardrails utilizing the Single-Turn Crescendo Attack (STCA)
Ted Kwartler, Nataliia Bagan, Ivan Banny, Alan Aqrawi, Arian Abbasi
Visual Adversarial Attack on Vision-Language Models for Autonomous Driving
Tianyuan Zhang, Lu Wang, Xinwei Zhang, Yitong Zhang, Boyi Jia, Siyuan Liang, Shengshan Hu, Qiang Fu, Aishan Liu, Xianglong Liu
Adversarial Training in Low-Label Regimes with Margin-Based Interpolation
Tian Ye, Rajgopal Kannan, Viktor Prasanna
Stealthy Multi-Task Adversarial Attacks
Jiacheng Guo, Tianyun Zhang, Lei Li, Haochen Yang, Hongkai Yu, Minghai Qin
Passive Deepfake Detection Across Multi-modalities: A Comprehensive Survey
Hong-Hanh Nguyen-Le, Van-Tuan Tran, Dinh-Thuc Nguyen, Nhien-An Le-Khac
Adversarial Bounding Boxes Generation (ABBG) Attack against Visual Object Trackers
Fatemeh Nourilenjan Nokabadi, Jean-Francois Lalonde, Christian Gagné
BadScan: An Architectural Backdoor Attack on Visual State Space Models
Om Suhas Deshmukh, Sankalp Nagaonkar, Achyut Mani Tripathi, Ashish Mishra
RED: Robust Environmental Design
Jinghan Yang
A Tunable Despeckling Neural Network Stabilized via Diffusion Equation
Yi Ran, Zhichang Guo, Jia Li, Yao Li, Martin Burger, Boying Wu
ExAL: An Exploration Enhanced Adversarial Learning Algorithm
A Vinil, Aneesh Sreevallabh Chivukula, Pranav Chintareddy
Chain of Attack: On the Robustness of Vision-Language Models Against Transfer-Based Adversarial Attacks
Peng Xie, Yequan Bie, Jianda Mao, Yangqiu Song, Yang Wang, Hao Chen, Kani Chen
Enhancing the Transferability of Adversarial Attacks on Face Recognition with Diverse Parameters Augmentation
Fengfan Zhou, Bangjie Yin, Hefei Ling, Qianyu Zhou, Wenxuan Wang
MUNBa: Machine Unlearning via Nash Bargaining
Jing Wu, Mehrtash Harandi
Steering Away from Harm: An Adaptive Approach to Defending Vision Language Model Against Jailbreaks
Han Wang, Gang Wang, Huan Zhang