Imperceptible Trigger

Imperceptible triggers are covert manipulations embedded in machine learning models, particularly deep neural networks, to induce unintended behavior without noticeably altering the model's performance on clean data. Current research focuses on developing increasingly stealthy triggers for various model types, including concept bottleneck models and self-supervised learning architectures, and on improving attack effectiveness while minimizing trigger detectability. This area is crucial for enhancing the security and trustworthiness of AI systems across diverse applications, from image classification and speech recognition to network security and reinforcement learning, by identifying and mitigating vulnerabilities to malicious manipulation.

Papers