Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
Papers
MOREL: Enhancing Adversarial Robustness through Multi-Objective Representation Learning
Sedjro Salomon Hotegni, Sebastian Peitz
On Using Certified Training towards Empirical Robustness
Alessandro De Palma, Serge Durand, Zakaria Chihani, François Terrier, Caterina Urban
Signal Adversarial Examples Generation for Signal Detection Network via White-Box Attack
Dongyang Li, Linyuan Wang, Guangwei Xiong, Bin Yan, Dekui Ma, Jinxian Peng
Ranking Over Scoring: Towards Reliable and Robust Automated Evaluation of LLM-Generated Medical Explanatory Arguments
Iker De la Iglesia, Iakes Goenaga, Johanna Ramirez-Romero, Jose Maria Villa-Gonzalez, Josu Goikoetxea, Ander Barrena
Navigating Threats: A Survey of Physical Adversarial Attacks on LiDAR Perception Systems in Autonomous Vehicles
Amira Guesmi, Muhammad Shafique
Robust LLM safeguarding via refusal feature adversarial training
Lei Yu, Virginie Do, Karen Hambardzumyan, Nicola Cancedda
Adversarial Examples for DNA Classification
Hyunwoo Yoo
Towards Robust Extractive Question Answering Models: Rethinking the Training Methodology
Son Quoc Tran, Matt Kretchmar
Nonideality-aware training makes memristive networks more robust to adversarial attacks
Dovydas Joksas, Luis Muñoz-González, Emil Lupu, Adnan Mehonic
Discerning the Chaos: Detecting Adversarial Perturbations while Disentangling Intentional from Unintentional Noises
Anubhooti Jain, Susim Roy, Kwanit Gupta, Mayank Vatsa, Richa Singh
MASKDROID: Robust Android Malware Detection with Masked Graph Representations
Jingnan Zheng, Jiaohao Liu, An Zhang, Jun Zeng, Ziqi Yang, Zhenkai Liang, Tat-Seng Chua
Evaluation of Security of ML-based Watermarking: Copy and Removal Attacks
Vitaliy Kinakh, Brian Pulfer, Yury Belousov, Pierre Fernandez, Teddy Furon, Slava Voloshynovskiy
Cross-Modality Attack Boosted by Gradient-Evolutionary Multiform Optimization
Yunpeng Gong, Qingyuan Zeng, Dejun Xu, Zhenzhong Wang, Min Jiang
Faithfulness and the Notion of Adversarial Sensitivity in NLP Explanations
Supriya Manna, Niladri Sett
Dark Miner: Defend against unsafe generation for text-to-image diffusion models
Zheling Meng, Bo Peng, Xiaochuan Jin, Yue Jiang, Jing Dong, Wei Wang, Tieniu Tan
Improving Fast Adversarial Training via Self-Knowledge Guidance
Chengze Jiang, Junkai Wang, Minjing Dong, Jie Gui, Xinli Shi, Yuan Cao, Yuan Yan Tang, James Tin-Yau Kwok
RED QUEEN: Safeguarding Large Language Models against Concealed Multi-Turn Jailbreaking
Yifan Jiang, Kriti Aggarwal, Tanmay Laud, Kashif Munir, Jay Pujara, Subhabrata Mukherjee