Adversarial Text
Adversarial text research focuses on creating and defending against text inputs designed to deceive natural language processing (NLP) models, often by subtly altering wording while maintaining semantic similarity to a human reader. Current research emphasizes developing more effective attack methods, particularly those leveraging multi-agent systems, reinforcement learning, and diffusion models, as well as improving defenses through techniques like adversarial training and noise augmentation. This field is crucial for enhancing the robustness and trustworthiness of NLP systems across diverse applications, from automated essay scoring to autonomous vehicle navigation and large language model safety.
Papers
May 24, 2023
May 16, 2023
April 21, 2023
April 8, 2023
March 31, 2023
March 24, 2023
March 2, 2023
February 22, 2023
February 12, 2023
February 5, 2023
February 3, 2023
February 2, 2023
November 13, 2022
November 12, 2022
November 5, 2022
November 1, 2022
October 28, 2022
October 22, 2022