Backdoor Attack
Backdoor attacks exploit vulnerabilities in machine learning models by embedding hidden triggers during training, causing the model to produce malicious outputs when the trigger is present. Current research focuses on developing and mitigating these attacks across various model architectures, including deep neural networks, vision transformers, graph neural networks, large language models, and spiking neural networks, with a particular emphasis on understanding attack mechanisms and developing robust defenses in federated learning and generative models. The significance of this research lies in ensuring the trustworthiness and security of increasingly prevalent machine learning systems across diverse applications, ranging from object detection and medical imaging to natural language processing and autonomous systems.
Papers
Fusing Pruned and Backdoored Models: Optimal Transport-based Data-free Backdoor Mitigation
Weilin Lin, Li Liu, Jianze Li, Hui Xiong
Defending Text-to-image Diffusion Models: Surprising Efficacy of Textual Perturbations Against Backdoor Attacks
Oscar Chew, Po-Yi Lu, Jayden Lin, Hsuan-Tien Lin
VFLIP: A Backdoor Defense for Vertical Federated Learning via Identification and Purification
Yungi Cho, Woorim Han, Miseon Yu, Younghan Lee, Ho Bae, Yunheung Paek
EmoAttack: Utilizing Emotional Voice Conversion for Speech Backdoor Attacks on Deep Speech Classification Models
Wenhan Yao, Zedong XingXiarun Chen, Jia Liu, yongqiang He, Weiping Wen