Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
Papers
Pre-trained Multiple Latent Variable Generative Models are good defenders against Adversarial Attacks
Dario Serez, Marco Cristani, Alessio Del Bue, Vittorio Murino, Pietro Morerio
Does Safety Training of LLMs Generalize to Semantically Related Natural Prompts?
Sravanti Addepalli, Yerram Varun, Arun Suggala, Karthikeyan Shanmugam, Prateek Jain
Testing Neural Network Verifiers: A Soundness Benchmark with Hidden Counterexamples
Xingjian Zhou, Hongji Xu, Andy Xu, Zhouxing Shi, Cho-Jui Hsieh, Huan Zhang
Out-of-Distribution Detection for Neurosymbolic Autonomous Cyber Agents
Ankita Samaddar, Nicholas Potteiger, Xenofon Koutsoukos
Gaussian Splatting Under Attack: Investigating Adversarial Noise in 3D Objects
Abdurrahman Zeybey, Mehmet Ergezer, Tommy Nguyen
Hijacking Vision-and-Language Navigation Agents with Adversarial Environmental Attacks
Zijiao Yang, Xiangxi Shi, Eric Slyman, Stefan Lee
Defending Against Diverse Attacks in Federated Learning Through Consensus-Based Bi-Level Optimization
Nicolás García Trillos, Aditya Kumar Akash, Sixu Li, Konstantin Riedl, Yuhua Zhu
Multi-Granularity Tibetan Textual Adversarial Attack Method Based on Masked Language Model
Xi Cao, Nuo Qun, Quzong Gesang, Yulei Zhu, Trashi Nyima
Fall Leaf Adversarial Attack on Traffic Sign Classification
Anthony Etim, Jakub Szefer
An indicator for effectiveness of text-to-image guardrails utilizing the Single-Turn Crescendo Attack (STCA)
Ted Kwartler, Nataliia Bagan, Ivan Banny, Alan Aqrawi, Arian Abbasi
Visual Adversarial Attack on Vision-Language Models for Autonomous Driving
Tianyuan Zhang, Lu Wang, Xinwei Zhang, Yitong Zhang, Boyi Jia, Siyuan Liang, Shengshan Hu, Qiang Fu, Aishan Liu, Xianglong Liu
Adversarial Training in Low-Label Regimes with Margin-Based Interpolation
Tian Ye, Rajgopal Kannan, Viktor Prasanna
Stealthy Multi-Task Adversarial Attacks
Jiacheng Guo, Tianyun Zhang, Lei Li, Haochen Yang, Hongkai Yu, Minghai Qin
Passive Deepfake Detection Across Multi-modalities: A Comprehensive Survey
Hong-Hanh Nguyen-Le, Van-Tuan Tran, Dinh-Thuc Nguyen, Nhien-An Le-Khac
Adversarial Bounding Boxes Generation (ABBG) Attack against Visual Object Trackers
Fatemeh Nourilenjan Nokabadi, Jean-Francois Lalonde, Christian Gagné