Imperceptible Attack

Imperceptible attacks exploit vulnerabilities in machine learning models by introducing subtle, human-undetectable perturbations to input data, causing misclassifications or biased outputs. Current research focuses on developing increasingly effective attacks across various modalities (images, text, audio) and model types (image classifiers, LLMs, speech recognition systems), often employing optimization algorithms like Frank-Wolfe or distributionally robust optimization to craft these perturbations. This research is crucial for evaluating the robustness of AI systems and informing the development of more secure and reliable models, with implications for applications ranging from facial recognition to autonomous systems.

Papers