Transferable Adversarial Attack
Transferable adversarial attacks aim to create perturbations in input data that fool multiple machine learning models, even those unseen during the attack's design. Current research focuses on improving the transferability of these attacks across diverse model architectures (including Vision Transformers, GANs, and diffusion models) and tasks (e.g., image classification, object detection, and language modeling), often employing techniques like gradient editing, contrastive learning, and frequency-domain manipulation. This research is crucial for evaluating the robustness of machine learning systems and informing the development of more secure and reliable AI applications, particularly in safety-critical domains.
Papers
November 14, 2024
October 30, 2024
October 26, 2024
October 24, 2024
October 9, 2024
September 6, 2024
August 22, 2024
July 30, 2024
July 21, 2024
July 18, 2024
June 8, 2024
May 10, 2024
May 9, 2024
April 30, 2024
April 12, 2024
April 3, 2024
February 26, 2024
February 16, 2024
February 2, 2024