Backdoor Attack
Backdoor attacks exploit vulnerabilities in machine learning models by embedding hidden triggers during training, causing the model to produce malicious outputs when the trigger is present. Current research focuses on developing and mitigating these attacks across various model architectures, including deep neural networks, vision transformers, graph neural networks, large language models, and spiking neural networks, with a particular emphasis on understanding attack mechanisms and developing robust defenses in federated learning and generative models. The significance of this research lies in ensuring the trustworthiness and security of increasingly prevalent machine learning systems across diverse applications, ranging from object detection and medical imaging to natural language processing and autonomous systems.
Papers
PatchBackdoor: Backdoor Attack against Deep Neural Networks without Model Modification
Yizhen Yuan, Rui Kong, Shenghao Xie, Yuanchun Li, Yunxin Liu
Protect Federated Learning Against Backdoor Attacks via Data-Free Trigger Generation
Yanxin Yang, Ming Hu, Yue Cao, Jun Xia, Yihao Huang, Yang Liu, Mingsong Chen
Improved Activation Clipping for Universal Backdoor Mitigation and Test-Time Detection
Hang Wang, Zhen Xiang, David J. Miller, George Kesidis
XGBD: Explanation-Guided Graph Backdoor Detection
Zihan Guan, Mengnan Du, Ninghao Liu
Breaking Speaker Recognition with PaddingBack
Zhe Ye, Diqun Yan, Li Dong, Kailai Shen
Backdoor Federated Learning by Poisoning Backdoor-Critical Layers
Haomin Zhuang, Mingxian Yu, Hao Wang, Yang Hua, Jian Li, Xu Yuan
Backdooring Instruction-Tuned Large Language Models with Virtual Prompt Injection
Jun Yan, Vikas Yadav, Shiyang Li, Lichang Chen, Zheng Tang, Hai Wang, Vijay Srinivasan, Xiang Ren, Hongxia Jin
BAGM: A Backdoor Attack for Manipulating Text-to-Image Generative Models
Jordan Vice, Naveed Akhtar, Richard Hartley, Ajmal Mian