Successful Adversarial Attack
Successful adversarial attacks exploit vulnerabilities in machine learning models by subtly altering inputs to cause misclassifications or undesired outputs. Current research focuses on developing more effective attack methods, particularly those that generate diverse and novel attacks across various model types, including large language models and image segmentation networks, often employing techniques like gradient-based optimization and reinforcement learning. Understanding and mitigating these attacks is crucial for ensuring the reliability and safety of AI systems across diverse applications, from autonomous vehicles to medical image analysis and online content moderation.
Papers
June 26, 2024
June 24, 2024
January 20, 2024
November 14, 2023
August 29, 2023
May 18, 2023
March 24, 2023
July 4, 2022
June 28, 2022
June 24, 2022
March 2, 2022
January 29, 2022
November 22, 2021