Sparse Perturbation
Sparse perturbation research focuses on understanding and mitigating the impact of small, targeted changes to data or model parameters. Current efforts concentrate on developing efficient algorithms for generating and defending against these sparse adversarial attacks, particularly within the context of image classification and language models, often employing techniques like projected gradient descent and sharpness-aware minimization with sparsified perturbations. This work is crucial for improving the robustness and security of machine learning systems, addressing vulnerabilities to malicious manipulations and enhancing the reliability of model explanations in various applications.
Papers
Improving Sharpness-Aware Minimization with Fisher Mask for Better Generalization on Language Models
Qihuang Zhong, Liang Ding, Li Shen, Peng Mi, Juhua Liu, Bo Du, Dacheng Tao
Make Sharpness-Aware Minimization Stronger: A Sparsified Perturbation Approach
Peng Mi, Li Shen, Tianhe Ren, Yiyi Zhou, Xiaoshuai Sun, Rongrong Ji, Dacheng Tao