Robust Defense
Robust defense in machine learning focuses on developing methods to protect models against various adversarial attacks, including backdoors, jailbreaks, data poisoning, and evasion attacks targeting different model architectures like LLMs and CNNs. Current research emphasizes developing defenses that are both effective against diverse attack strategies and efficient, addressing challenges like resource constraints and privacy concerns through techniques such as randomized smoothing, gradient masking, and reinforcement learning-based approaches. These advancements are crucial for ensuring the reliability and trustworthiness of AI systems across diverse applications, from autonomous driving to medical diagnosis.
Papers
No Free Lunch for Defending Against Prefilling Attack by In-Context Learning
Zhiyu Xue, Guangliang Liu, Bocheng Chen, Kristen Marie Johnson, Ramtin Pedarsani
BiCert: A Bilinear Mixed Integer Programming Formulation for Precise Certified Bounds Against Data Poisoning Attacks
Tobias Lorenz, Marta Kwiatkowska, Mario Fritz