Manifold Attack

Manifold attacks exploit the geometric structure of data to generate adversarial examples—inputs subtly altered to fool machine learning models while remaining perceptually similar to legitimate data points. Current research focuses on distinguishing between "on-manifold" attacks, which stay within the natural data distribution, and "off-manifold" attacks that deviate from it, employing techniques like projected gradient descent and manifold approximations (e.g., using diffusion maps) to generate and defend against these attacks. Understanding and mitigating the vulnerability of models to these attacks is crucial for improving the robustness and reliability of machine learning systems across various applications, including image classification and natural language processing.

Papers