White Box Adversarial Attack

White-box adversarial attacks aim to evaluate and exploit the vulnerabilities of machine learning models by crafting malicious inputs designed to cause misclassification, leveraging full knowledge of the model's architecture and parameters. Current research focuses on developing and analyzing these attacks across various model types, including convolutional neural networks, transformers, and hypergraph neural networks, using algorithms like Projected Gradient Descent and its variants. Understanding the susceptibility of these models to such attacks is crucial for improving their robustness and ensuring the reliable deployment of machine learning in safety-critical applications like autonomous vehicles and healthcare.

Papers