Adversarial Capability

Adversarial capability research explores how to create and defend against malicious inputs designed to deceive machine learning models. Current efforts focus on developing more effective attack methods, including those employing reinforcement learning, disentangled feature spaces, and techniques tailored to specific data types like tabular data and hardware power traces. This research is crucial for improving the robustness and security of machine learning systems across various applications, from malware detection to safety-critical systems, by identifying vulnerabilities and developing more resilient models. The ultimate goal is to create models that are not only accurate but also resistant to manipulation.

Papers