Adversarial Search
Adversarial search explores methods for finding inputs that intentionally mislead machine learning models, revealing vulnerabilities and improving robustness. Current research focuses on developing efficient attack algorithms, such as beam search methods, and robust defenses, including techniques leveraging uncertainty estimation and classical machine learning models for verification. This field is crucial for enhancing the security and reliability of AI systems across diverse applications, from image recognition and natural language processing to autonomous systems and search engines. The development of both effective attacks and defenses is driving improvements in model design and evaluation methodologies.
Papers
Discovering Failure Modes of Text-guided Diffusion Models via Adversarial Search
Qihao Liu, Adam Kortylewski, Yutong Bai, Song Bai, Alan Yuille
Adversarial-Aware Deep Learning System based on a Secondary Classical Machine Learning Verification Approach
Mohammed Alkhowaiter, Hisham Kholidy, Mnassar Alyami, Abdulmajeed Alghamdi, Cliff Zou