Perturbation Attack
Perturbation attacks involve subtly altering inputs to machine learning models, particularly deep neural networks, to cause misclassification or inaccurate outputs. Current research focuses on developing increasingly effective attack strategies, including those targeting specific model architectures (e.g., vision foundation models, no-reference image quality metrics) and employing diverse perturbation methods (e.g., pixel-wise adjustments under various norms, universal perturbations). This research is crucial for evaluating and improving the robustness of machine learning systems across various applications, from autonomous driving to medical image analysis, where the reliability of model predictions is paramount. Understanding and mitigating these attacks is essential for ensuring the safe and trustworthy deployment of AI systems.