Adversarial Instance
Adversarial instances are subtly modified inputs designed to fool machine learning models, primarily deep neural networks (DNNs), into making incorrect predictions. Current research focuses on generating these instances using evolutionary algorithms and adversarial training techniques, analyzing their impact on model robustness and fairness across various architectures including recurrent and diffusion models, and developing methods to detect and mitigate their effects. Understanding and addressing adversarial instances is crucial for improving the reliability and trustworthiness of AI systems in safety-critical applications and advancing the broader field of machine learning.
18papers
Papers
January 19, 2025
June 24, 2024
January 12, 2024
November 14, 2023
August 11, 2023
December 13, 2022
August 29, 2022
August 17, 2022
May 2, 2022