Adversary Knowledge

Adversary knowledge, in the context of machine learning, explores how an attacker's understanding of a target model influences the effectiveness of adversarial attacks—malicious inputs designed to cause misclassification or other failures. Current research focuses on characterizing the impact of different levels of adversary knowledge (e.g., white-box vs. black-box access) across various model architectures, including deep neural networks, graph neural networks, and reinforcement learning agents, and developing frameworks to systematically analyze these attacks. Understanding adversary knowledge is crucial for improving the robustness and security of machine learning systems in diverse applications, from network security to autonomous systems and data privacy.

Papers