Clean Label Backdoor Attack
Clean-label backdoor attacks are a stealthy form of data poisoning where attackers manipulate training data to control model predictions without altering the labels, making detection challenging. Current research focuses on developing more effective attack strategies, including those leveraging self-supervised learning and language models, as well as exploring defenses against these attacks, such as density-based clustering and label smoothing techniques, across various model architectures and data modalities (image, text, video). The significance lies in the potential for widespread malicious manipulation of machine learning models in security-critical applications, highlighting the urgent need for robust defenses and improved model security.