Transferable Adversarial Attack
Transferable adversarial attacks aim to create perturbations in input data that fool multiple machine learning models, even those unseen during the attack's design. Current research focuses on improving the transferability of these attacks across diverse model architectures (including Vision Transformers, GANs, and diffusion models) and tasks (e.g., image classification, object detection, and language modeling), often employing techniques like gradient editing, contrastive learning, and frequency-domain manipulation. This research is crucial for evaluating the robustness of machine learning systems and informing the development of more secure and reliable AI applications, particularly in safety-critical domains.
Papers
February 1, 2024
January 11, 2024
December 11, 2023
October 19, 2023
October 1, 2023
September 19, 2023
August 30, 2023
July 27, 2023
July 6, 2023
May 14, 2023
March 28, 2023
March 22, 2023
March 7, 2023
November 16, 2022
October 13, 2022
October 9, 2022
September 24, 2022
August 11, 2022
July 18, 2022