Unrestricted Adversarial

Unrestricted adversarial attacks aim to create maliciously perturbed data that fool machine learning models while appearing natural to humans, unlike traditional attacks limited by specific perturbation norms. Current research focuses on generating these attacks using diffusion models and generative adversarial networks (GANs), often incorporating techniques like latent space manipulation, semantic guidance from large language models, and recursive token merging for improved realism and transferability across different models. This research is crucial for evaluating the robustness of machine learning systems in real-world scenarios and informing the development of more resilient models and defenses against sophisticated attacks.

Papers