Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
Papers
Towards Resilient and Efficient LLMs: A Comparative Study of Efficiency, Performance, and Adversarial Robustness
Xiaojing Fan, Chunliang Tao
Ensemble everything everywhere: Multi-scale aggregation for adversarial robustness
Stanislav Fort, Balaji Lakshminarayanan
Constructing Adversarial Examples for Vertical Federated Learning: Optimal Client Corruption through Multi-Armed Bandit
Duanyi Yao, Songze Li, Ye Xue, Jin Liu
EdgeShield: A Universal and Efficient Edge Computing Framework for Robust AI
Duo Zhong, Bojing Li, Xiang Chen, Chenchen Liu
LaFA: Latent Feature Attacks on Non-negative Matrix Factorization
Minh Vu, Ben Nebgen, Erik Skau, Geigh Zollicoffer, Juan Castorena, Kim Rasmussen, Boian Alexandrov, Manish Bhattarai
Enhancing Output Diversity Improves Conjugate Gradient-based Adversarial Attacks
Keiichiro Yamamura, Issa Oe, Hiroki Ishikura, Katsuki Fujisawa
Simple Perturbations Subvert Ethereum Phishing Transactions Detection: An Empirical Analysis
Ahod Alghureid, David Mohaisen
Attacks and Defenses for Generative Diffusion Models: A Comprehensive Survey
Vu Tuan Truong, Luan Ba Dang, Long Bao Le
Sample-agnostic Adversarial Perturbation for Vision-Language Pre-training Models
Haonan Zheng, Wen Jiang, Xinyang Deng, Wenrui Li
Pre-trained Encoder Inference: Revealing Upstream Encoders In Downstream Machine Learning Services
Shaopeng Fu, Xuexue Sun, Ke Qing, Tianhang Zheng, Di Wang
SEAS: Self-Evolving Adversarial Safety Optimization for Large Language Models
Muxi Diao, Rumei Li, Shiyang Liu, Guogang Liao, Jingang Wang, Xunliang Cai, Weiran Xu
On the Robustness of Malware Detectors to Adversarial Samples
Muhammad Salman, Benjamin Zi Hao Zhao, Hassan Jameel Asghar, Muhammad Ikram, Sidharth Kaushik, Mohamed Ali Kaafar
AdvQDet: Detecting Query-Based Adversarial Attacks with Adversarial Contrastive Prompt Tuning
Xin Wang, Kai Chen, Xingjun Ma, Zhineng Chen, Jingjing Chen, Yu-Gang Jiang
Top K Enhanced Reinforcement Learning Attacks on Heterogeneous Graph Node Classification
Honglin Gao, Gaoxi Xiao
A Survey and Evaluation of Adversarial Attacks for Object Detection
Khoi Nguyen Tiet Nguyen, Wenyu Zhang, Kangkang Lu, Yuhuan Wu, Xingjian Zheng, Hui Li Tan, Liangli Zhen
ALIF: Low-Cost Adversarial Audio Attacks on Black-Box Speech Platforms using Linguistic Features
Peng Cheng, Yuwei Wang, Peng Huang, Zhongjie Ba, Xiaodong Lin, Feng Lin, Li Lu, Kui Ren
Joint Universal Adversarial Perturbations with Interpretations
Liang-bo Ning, Zeyu Dai, Wenqi Fan, Jingran Su, Chao Pan, Luning Wang, Qing Li
Downstream Transfer Attack: Adversarial Attacks on Downstream Models with Pre-trained Vision Transformers
Weijie Zheng, Xingjun Ma, Hanxun Huang, Zuxuan Wu, Yu-Gang Jiang
Guardians of Image Quality: Benchmarking Defenses Against Adversarial Attacks on Image Quality Metrics
Alexander Gushchin, Khaled Abud, Georgii Bychkov, Ekaterina Shumitskaya, Anna Chistyakova, Sergey Lavrushkin, Bader Rasheed, Kirill Malyshev, Dmitriy Vatolin, Anastasia Antsiferova
Assessing Robustness of Machine Learning Models using Covariate Perturbations
Arun Prakash R, Anwesha Bhattacharyya, Joel Vaughan, Vijayan N. Nair