Adversarial Instance
Adversarial instances are subtly modified inputs designed to fool machine learning models, primarily deep neural networks (DNNs), into making incorrect predictions. Current research focuses on generating these instances using evolutionary algorithms and adversarial training techniques, analyzing their impact on model robustness and fairness across various architectures including recurrent and diffusion models, and developing methods to detect and mitigate their effects. Understanding and addressing adversarial instances is crucial for improving the reliability and trustworthiness of AI systems in safety-critical applications and advancing the broader field of machine learning.
Papers
June 24, 2024
May 1, 2024
April 1, 2024
January 12, 2024
November 14, 2023
August 11, 2023
August 3, 2023
May 18, 2023
March 13, 2023
December 13, 2022
August 29, 2022
August 17, 2022
June 2, 2022
May 20, 2022
May 2, 2022
January 25, 2022
December 14, 2021