Adversarial Instance

Adversarial instances are subtly modified inputs designed to fool machine learning models, primarily deep neural networks (DNNs), into making incorrect predictions. Current research focuses on generating these instances using evolutionary algorithms and adversarial training techniques, analyzing their impact on model robustness and fairness across various architectures including recurrent and diffusion models, and developing methods to detect and mitigate their effects. Understanding and addressing adversarial instances is crucial for improving the reliability and trustworthiness of AI systems in safety-critical applications and advancing the broader field of machine learning.

Papers