Poisoning Attack
Poisoning attacks target machine learning models by injecting malicious data into the training process, aiming to degrade model performance or introduce backdoors. Current research focuses on developing sophisticated poisoning strategies for various model architectures, including decision trees, neural networks, and recommender systems, and exploring defenses such as robust aggregation techniques and anomaly detection methods in federated learning settings. Understanding and mitigating these attacks is crucial for ensuring the reliability and security of machine learning systems across diverse applications, from autonomous driving to financial services. The ongoing development of both more effective attacks and more robust defenses highlights the importance of this area of research.
Papers
Fragile Giants: Understanding the Susceptibility of Models to Subpopulation Attacks
Isha Gupta, Hidde Lycklama, Emanuel Opel, Evan Rose, Anwar Hithnawi
PoisonBench: Assessing Large Language Model Vulnerability to Data Poisoning
Tingchen Fu, Mrinank Sharma, Philip Torr, Shay B. Cohen, David Krueger, Fazl Barez