Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
Papers
Towards Robust and Accurate Stability Estimation of Local Surrogate Models in Text-based Explainable AI
Christopher Burger, Charles Walter, Thai Le, Lingwei Chen
Detecting and Mitigating Adversarial Attacks on Deep Learning-Based MRI Reconstruction Without Any Retraining
Mahdi Saberi, Chi Zhang, Mehmet Akcakaya
Adaptive Meta-learning-based Adversarial Training for Robust Automatic Modulation Classification
Amirmohammad Bamdad, Ali Owfi, Fatemeh Afghah
Safeguarding Large Language Models in Real-time with Tunable Safety-Performance Trade-offs
Joao Fonseca, Andrew Bell, Julia Stoyanovich
Image-based Multimodal Models as Intruders: Transferable Multimodal Attacks on Video-based MLLMs
Linhao Huang, Xue Jiang, Zhiqiang Wang, Wentao Mo, Xi Xiao, Bo Han, Yongjie Yin, Feng Zheng
Towards Adversarially Robust Deep Metric Learning
Xiaopeng Ke
Federated Hybrid Training and Self-Adversarial Distillation: Towards Robust Edge Networks
Yu Qiao, Apurba Adhikary, Kitae Kim, Eui-Nam Huh, Zhu Han, Choong Seon Hong
Imperceptible Adversarial Attacks on Point Clouds Guided by Point-to-Surface Field
Keke Tang, Weiyao Ke, Weilong Peng, Xiaofei Wang, Ziyong Du, Zhize Wu, Peican Zhu, Zhihong Tian
Bridging Interpretability and Robustness Using LIME-Guided Model Refinement
Navid Nayyem, Abdullah Rakin, Longwei Wang
Distortion-Aware Adversarial Attacks on Bounding Boxes of Object Detectors
Pham Phuc, Son Vuong, Khang Nguyen, Tuan Dang
Protective Perturbations against Unauthorized Data Usage in Diffusion-based Image Generation
Sen Peng, Jijia Yang, Mingyue Wang, Jianfei He, Xiaohua Jia
Evaluating the Adversarial Robustness of Detection Transformers
Amirhossein Nazeri, Chunheng Zhao, Pierluigi Pisu