Backdoor Data
Backdoor data poisoning attacks compromise machine learning models by subtly embedding malicious triggers within training data, causing the model to produce incorrect outputs under specific conditions while appearing normal otherwise. Current research focuses on detecting these poisoned datasets, often leveraging prediction uncertainty analysis or geometric data properties to identify suspicious samples, and developing robust training methods that mitigate the impact of backdoor triggers, including techniques that focus on specific model layers or frequency space manipulations. Understanding and defending against backdoor attacks is crucial for ensuring the reliability and security of machine learning systems across various applications, from large language models to federated learning environments.