Poisoning Attack
Poisoning attacks target machine learning models by injecting malicious data into the training process, aiming to degrade model performance or introduce backdoors. Current research focuses on developing sophisticated poisoning strategies for various model architectures, including decision trees, neural networks, and recommender systems, and exploring defenses such as robust aggregation techniques and anomaly detection methods in federated learning settings. Understanding and mitigating these attacks is crucial for ensuring the reliability and security of machine learning systems across diverse applications, from autonomous driving to financial services. The ongoing development of both more effective attacks and more robust defenses highlights the importance of this area of research.
Papers
Manipulating Recommender Systems: A Survey of Poisoning Attacks and Countermeasures
Thanh Toan Nguyen, Quoc Viet Hung Nguyen, Thanh Tam Nguyen, Thanh Trung Huynh, Thanh Thi Nguyen, Matthias Weidlich, Hongzhi Yin
Talk Too Much: Poisoning Large Language Models under Token Limit
Jiaming He, Wenbo Jiang, Guanyu Hou, Wenshu Fan, Rui Zhang, Hongwei Li