Adversarial Information
Adversarial information research focuses on understanding and mitigating the impact of malicious or misleading information, particularly in the context of large language models and online environments. Current research explores methods to improve model robustness against adversarial attacks, including developing new metrics for evaluating model resilience and employing techniques like information bottleneck principles and adversarial training to enhance resistance to manipulation. This work is crucial for building more reliable AI systems and for combating the spread of misinformation, with implications for election integrity, public opinion, and the trustworthiness of online information.
Papers
October 15, 2024
October 10, 2024
August 4, 2024
May 6, 2024
February 12, 2024
December 4, 2023
October 12, 2023
May 23, 2023
November 1, 2022
May 3, 2022
February 22, 2022
January 21, 2022