Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
Papers
Well, that escalated quickly: The Single-Turn Crescendo Attack (STCA)
Alan Aqrawi
OpenFact at CheckThat! 2024: Combining Multiple Attack Methods for Effective Adversarial Text Generation
Włodzimierz Lewoniewski, Piotr Stolarski, Milena Stróżyna, Elzbieta Lewańska, Aleksandra Wojewoda, Ewelina Księżniak, Marcin Sawiński
AdvSecureNet: A Python Toolkit for Adversarial Machine Learning
Melih Catal, Manuel Günther
Adversarial Attacks on Machine Learning-Aided Visualizations
Takanori Fujiwara, Kostiantyn Kucher, Junpeng Wang, Rafael M. Martins, Andreas Kerren, Anders Ynnerman
Exact Recovery Guarantees for Parameterized Non-linear System Identification Problem under Adversarial Attacks
Haixiang Zhang, Baturalp Yalcin, Javad Lavaei, Eduardo D. Sontag
Discovery of False Data Injection Schemes on Frequency Controllers with Reinforcement Learning
Romesh Prasad, Malik Hassanaly, Xiangyu Zhang, Abhijeet Sahu
Multi-modal Adversarial Training for Zero-Shot Voice Cloning
John Janiczek, Dading Chong, Dongyang Dai, Arlo Faria, Chao Wang, Tao Wang, Yuzong Liu
Evaluating Model Robustness Using Adaptive Sparse L0 Regularization
Weiyou Liu, Zhenyang Li, Weitong Chen
Certified Causal Defense with Generalizable Robustness
Yiran Qiao, Yu Yin, Chen Chen, Jing Ma
Adversarial Attacks and Defenses in Multivariate Time-Series Forecasting for Smart and Connected Infrastructures
Pooja Krishan, Rohan Mohapatra, Saptarshi Sengupta
Feedback-based Modal Mutual Search for Attacking Vision-Language Pre-training Models
Renhua Ding, Xinze Zhang, Xiao Yang, Kun He
TART: Boosting Clean Accuracy Through Tangent Direction Guided Adversarial Training
Bongsoo Yi, Rongjie Lai, Yao Li