Sparse Perturbation

Sparse perturbation research focuses on understanding and mitigating the impact of small, targeted changes to data or model parameters. Current efforts concentrate on developing efficient algorithms for generating and defending against these sparse adversarial attacks, particularly within the context of image classification and language models, often employing techniques like projected gradient descent and sharpness-aware minimization with sparsified perturbations. This work is crucial for improving the robustness and security of machine learning systems, addressing vulnerabilities to malicious manipulations and enhancing the reliability of model explanations in various applications.

Papers