Adversarial Attack
Adversarial attacks aim to deceive machine learning models by subtly altering input data, causing misclassifications or other erroneous outputs. Current research focuses on developing more robust models and detection methods, exploring various attack strategies across different model architectures (including vision transformers, recurrent neural networks, and graph neural networks) and data types (images, text, signals, and tabular data). Understanding and mitigating these attacks is crucial for ensuring the reliability and security of AI systems in diverse applications, from autonomous vehicles to medical diagnosis and cybersecurity.
1813papers
Papers - Page 10
January 27, 2025
Towards Safe AI Clinicians: A Comprehensive Study on Large Language Model Jailbreaking in Healthcare
Towards Robust Stability Prediction in Smart Grids: GAN-based Approach under Data Constraints and Adversarial Challenges
The Relationship Between Network Similarity and Transferability of Adversarial Attacks
The TIP of the Iceberg: Revealing a Hidden Class of Task-In-Prompt Adversarial Attacks on LLMs
January 25, 2025
January 23, 2025
January 21, 2025
Robustness of Selected Learning Models under Label-Flipping Attack
With Great Backbones Comes Great Adversarial Transferability
Heterogeneous Multi-Player Multi-Armed Bandits Robust To Adversarial Attacks
Transferable Adversarial Attacks on Audio Deepfake Detection
Enhancing Adversarial Transferability via Component-Wise Transformation
January 20, 2025