Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
Papers
The Impacts of Unanswerable Questions on the Robustness of Machine Reading Comprehension Models
Son Quoc Tran, Phong Nguyen-Thuan Do, Uyen Le, Matt Kretchmar
Reverse engineering adversarial attacks with fingerprints from adversarial examples
David Aaron Nicholson, Vincent Emanuele
Robust Linear Regression: Gradient-descent, Early-stopping, and Beyond
Meyer Scetbon, Elvis Dohmatob
Inference Time Evidences of Adversarial Attacks for Forensic on Transformers
Hugo Lemarchant, Liangzi Li, Yiming Qian, Yuta Nakashima, Hajime Nagahara
Affinity Uncertainty-based Hard Negative Mining in Graph Contrastive Learning
Chaoxi Niu, Guansong Pang, Ling Chen
Towards Adversarial Realism and Robust Learning for IoT Intrusion Detection and Classification
João Vitorino, Isabel Praça, Eva Maia
Language-Driven Anchors for Zero-Shot Adversarial Robustness
Xiao Li, Wei Zhang, Yining Liu, Zhanhao Hu, Bo Zhang, Xiaolin Hu
On the Efficacy of Metrics to Describe Adversarial Attacks
Tommaso Puccetti, Tommaso Zoppi, Andrea Ceccarelli
Improving Adversarial Transferability with Scheduled Step Size and Dual Example
Zeliang Zhang, Peihan Liu, Xiaosen Wang, Chenliang Xu
Identifying Adversarially Attackable and Robust Samples
Vyas Raina, Mark Gales
Adversarial Attacks on Adversarial Bandits
Yuzhe Ma, Zhijin Zhou
Node Injection for Class-specific Network Poisoning
Ansh Kumar Sharma, Rahul Kukreja, Mayank Kharbanda, Tanmoy Chakraborty
Selecting Models based on the Risk of Damage Caused by Adversarial Attacks
Jona Klemenc, Holger Trittenbach
Semantic Adversarial Attacks on Face Recognition through Significant Attributes
Yasmeen M. Khedr, Yifeng Xiong, Kun He
Certified Invertibility in Neural Networks via Mixed-Integer Programming
Tianqi Cui, Thomas Bertalan, George J. Pappas, Manfred Morari, Ioannis G. Kevrekidis, Mahyar Fazlyab
Learning to Unlearn: Instance-wise Unlearning for Pre-trained Classifiers
Sungmin Cha, Sungjun Cho, Dasol Hwang, Honglak Lee, Taesup Moon, Moontae Lee
Adapting Step-size: A Unified Perspective to Analyze and Improve Gradient-based Methods for Adversarial Attacks
Wei Tao, Lei Bao, Sheng Long, Gaowei Wu, Qing Tao
Targeted Attacks on Timeseries Forecasting
Yuvaraj Govindarajulu, Avinash Amballa, Pavan Kulkarni, Manojkumar Parmar