Malicious Training

Malicious training, a significant threat to machine learning, involves manipulating training data to compromise the integrity and performance of trained models. Current research focuses on developing robust defenses against various attack vectors, including data poisoning (introducing malicious samples) and adversarial examples (subtly altered inputs), employing techniques like temporal analysis of data timestamps, influence graphs to identify malicious samples, and data augmentation to improve model robustness. Understanding and mitigating these attacks is crucial for ensuring the reliability and security of machine learning systems across diverse applications, from malware detection to critical infrastructure.

Papers