Adversarial Text
Adversarial text research focuses on creating and defending against text inputs designed to deceive natural language processing (NLP) models, often by subtly altering wording while maintaining semantic similarity to a human reader. Current research emphasizes developing more effective attack methods, particularly those leveraging multi-agent systems, reinforcement learning, and diffusion models, as well as improving defenses through techniques like adversarial training and noise augmentation. This field is crucial for enhancing the robustness and trustworthiness of NLP systems across diverse applications, from automated essay scoring to autonomous vehicle navigation and large language model safety.
Papers
February 5, 2024
February 1, 2024
January 22, 2024
January 18, 2024
December 12, 2023
December 8, 2023
December 3, 2023
November 14, 2023
October 31, 2023
October 29, 2023
October 25, 2023
October 12, 2023
September 30, 2023
July 27, 2023
June 14, 2023
June 9, 2023
June 7, 2023
June 6, 2023
June 2, 2023