Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
Papers
Friendly Attacks to Improve Channel Coding Reliability
Anastasiia Kurmukova, Deniz Gunduz
Sparse and Transferable Universal Singular Vectors Attack
Kseniia Kuvshinova, Olga Tsymboi, Ivan Oseledets
Respect the model: Fine-grained and Robust Explanation with Sharing Ratio Decomposition
Sangyu Han, Yearim Kim, Nojun Kwak
How Robust Are Energy-Based Models Trained With Equilibrium Propagation?
Siddharth Mansingh, Michal Kucer, Garrett Kenyon, Juston Moore, Michael Teti
Finding a Needle in the Adversarial Haystack: A Targeted Paraphrasing Approach For Uncovering Edge Cases with Minimal Distribution Distortion
Aly M. Kassem, Sherif Saad