Adversarial Change
Adversarial change research investigates how machine learning models react to intentionally malicious modifications of their inputs or training data, aiming to improve model robustness and trustworthiness. Current research focuses on developing methods to evaluate and enhance model resilience against these attacks across various domains, including image classification, natural language processing, and reinforcement learning, employing techniques like adversarial training and bi-level optimization. This work is crucial for ensuring the reliability and safety of AI systems deployed in high-stakes applications, where vulnerabilities to adversarial manipulation can have significant consequences.
Papers
October 2, 2024
September 26, 2024
July 15, 2024
June 6, 2024
March 7, 2024
January 23, 2024
December 17, 2023
May 18, 2023
January 30, 2023
December 18, 2022