Adversarial Data
Adversarial data, encompassing maliciously perturbed inputs designed to mislead machine learning models, poses a significant threat to the reliability of AI systems. Current research focuses on developing robust models through techniques like adversarial training (incorporating adversarial examples during training), and novel detection methods that identify adversarial instances based on distributional discrepancies or feature analysis, often employing diffusion models or ensemble approaches. This field is crucial for ensuring the trustworthiness and security of AI applications across diverse domains, from medical diagnosis and autonomous driving to natural language processing, where adversarial attacks can have serious consequences.
Papers
September 18, 2024
August 23, 2024
August 2, 2024
August 1, 2024
June 18, 2024
May 25, 2024
May 23, 2024
April 9, 2024
March 5, 2024
February 28, 2024
February 13, 2024
February 3, 2024
August 1, 2023
June 22, 2023
June 2, 2023
May 26, 2023
May 25, 2023
April 30, 2023
April 21, 2023