Poisoned Data

Poisoned data, the malicious injection of corrupted samples into training datasets, poses a significant threat to the reliability and security of machine learning models. Current research focuses on developing robust defenses, including methods that leverage self-supervised learning, filter poisoned samples based on backdoor attack characteristics, and selectively "unlearn" poisoned data from already trained models. These efforts are crucial for ensuring the trustworthiness of machine learning systems across various applications, particularly in sensitive domains like healthcare and autonomous driving, where model robustness is paramount.

Papers