Backdoor Threat
Backdoor attacks exploit vulnerabilities in machine learning models, causing them to produce specific, malicious outputs when presented with a hidden "trigger," while appearing normal otherwise. Current research focuses on understanding and mitigating these attacks across various model architectures, including those used in federated learning and natural language processing, with a particular emphasis on identifying effective defenses and developing robust evaluation metrics. The significance of this research lies in ensuring the trustworthiness and security of machine learning systems deployed in critical applications, where compromised models could have severe consequences.
Papers
June 10, 2024
April 17, 2024
February 21, 2024
November 25, 2023
August 31, 2023
May 16, 2023