Imperceptible Attack
Imperceptible attacks exploit vulnerabilities in machine learning models by introducing subtle, human-undetectable perturbations to input data, causing misclassifications or biased outputs. Current research focuses on developing increasingly effective attacks across various modalities (images, text, audio) and model types (image classifiers, LLMs, speech recognition systems), often employing optimization algorithms like Frank-Wolfe or distributionally robust optimization to craft these perturbations. This research is crucial for evaluating the robustness of AI systems and informing the development of more secure and reliable models, with implications for applications ranging from facial recognition to autonomous systems.
Papers
January 9, 2025
December 15, 2024
July 11, 2024
June 7, 2024
April 9, 2024
February 14, 2024
November 27, 2023
October 12, 2023
July 25, 2023
June 1, 2023
December 14, 2022
October 15, 2022
July 20, 2022
July 17, 2022
June 4, 2022
March 17, 2022
March 3, 2022
November 30, 2021