Adversarial Example Detection
Adversarial example detection aims to identify inputs maliciously crafted to fool machine learning models, a critical challenge for deploying these models in safety-critical applications. Current research focuses on improving the generalization ability of detectors across various attack types, exploring techniques like data reconstruction using variational autoencoders and leveraging features such as high-frequency image differences or even sentiment analysis of model internal representations. Effective adversarial example detection is crucial for enhancing the robustness and trustworthiness of machine learning systems, impacting fields ranging from autonomous driving to cybersecurity.
Papers
April 19, 2024
June 3, 2023
May 8, 2023
May 3, 2023
November 10, 2022
August 31, 2022
June 30, 2022