Textual Adversarial Attack
Textual adversarial attacks involve subtly altering text inputs to mislead natural language processing (NLP) models, primarily focusing on evaluating and improving model robustness. Current research emphasizes developing more effective attack methods, often leveraging gradient-based optimization and incorporating semantic similarity constraints within models like BERT, as well as creating stronger defenses through techniques such as adversarial training and randomized smoothing. This field is crucial for ensuring the reliability and security of NLP systems across various applications, from text classification to machine translation, by identifying and mitigating vulnerabilities to malicious manipulation.
Papers
October 11, 2024
August 19, 2024
July 29, 2024
June 18, 2024
May 25, 2024
April 8, 2024
February 29, 2024
February 5, 2024
January 23, 2024
January 7, 2024
November 29, 2023
October 21, 2023
July 31, 2023
July 24, 2023
July 4, 2023
June 27, 2023
June 7, 2023
February 18, 2023
February 12, 2023