Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
Papers
Certification of Speaker Recognition Models to Additive Perturbations
Dmitrii Korzh, Elvir Karimov, Mikhail Pautov, Oleg Y. Rogov, Ivan Oseledets
Why You Should Not Trust Interpretations in Machine Learning: Adversarial Attacks on Partial Dependence Plots
Xi Xin, Giles Hooker, Fei Huang
Towards Quantitative Evaluation of Explainable AI Methods for Deepfake Detection
Konstantinos Tsigos, Evlampios Apostolidis, Spyridon Baxevanakis, Symeon Papadopoulos, Vasileios Mezaris
A Systematic Evaluation of Adversarial Attacks against Speech Emotion Recognition Models
Nicolas Facchinetti, Federico Simonetta, Stavros Ntalampiras
Bounding the Expected Robustness of Graph Neural Networks Subject to Node Feature Attacks
Yassine Abbahaddou, Sofiane Ennadir, Johannes F. Lutzeyer, Michalis Vazirgiannis, Henrik Boström
Attacking Bayes: On the Adversarial Robustness of Bayesian Neural Networks
Yunzhen Feng, Tim G. J. Rudner, Nikolaos Tsilivis, Julia Kempe
Generating Minimalist Adversarial Perturbations to Test Object-Detection Models: An Adaptive Multi-Metric Evolutionary Search Approach
Cristopher McIntyre-Garcia, Adrien Heymans, Beril Borali, Won-Sook Lee, Shiva Nejati
Don't Say No: Jailbreaking LLM by Suppressing Refusal
Yukai Zhou, Zhijie Huang, Feiyang Lu, Zhan Qin, Wenjie Wang
A Comparative Analysis of Adversarial Robustness for Quantum and Classical Machine Learning Models
Maximilian Wendlinger, Kilian Tscharke, Pascal Debus
A Survey on Intermediate Fusion Methods for Collaborative Perception Categorized by Real World Challenges
Melih Yazgan, Thomas Graf, Min Liu, Tobias Fleck, J. Marius Zoellner
Steal Now and Attack Later: Evaluating Robustness of Object Detection against Black-box Adversarial Attacks
Erh-Chung Chen, Pin-Yu Chen, I-Hsin Chung, Che-Rung Lee
An Empirical Study of Aegis
Daniel Saragih, Paridhi Goel, Tejas Balaji, Alyssa Li
MISLEAD: Manipulating Importance of Selected features for Learning Epsilon in Evasion Attack Deception
Vidit Khazanchi, Pavan Kulkarni, Yuvaraj Govindarajulu, Manojkumar Parmar
Formal Verification of Graph Convolutional Networks with Uncertain Node Features and Uncertain Graph Structure
Tobias Ladner, Michael Eichelbeck, Matthias Althoff
DIP-Watermark: A Double Identity Protection Method Based on Robust Adversarial Watermark
Yunming Zhang, Dengpan Ye, Caiyun Xie, Sipeng Shen, Ziyi Liu, Jiacheng Deng, Long Tang