Resilient Backdoor
Resilient backdoor attacks aim to embed malicious functionality into machine learning models, making them vulnerable to manipulation by injecting specific triggers that cause misclassification. Current research focuses on developing more robust backdoor techniques that resist detection and mitigation methods, including those leveraging continual learning, architectural modifications, and sophisticated trigger generation strategies designed to evade detection. This area is crucial because the persistence and difficulty of detecting these backdoors pose significant security risks to various machine learning applications, demanding the development of more effective defenses.
Papers
July 20, 2024
May 30, 2024
March 25, 2024
February 10, 2024
November 28, 2023
April 25, 2023
April 20, 2023
March 1, 2023
June 15, 2022
June 12, 2022