Content Moderator
Content moderation, the process of screening and managing user-generated content online, aims to maintain safe and productive online environments by removing harmful or inappropriate material. Current research focuses on improving automated moderation systems, employing techniques like natural language processing (NLP) and large language models (LLMs) to identify violations of community rules and policies, often incorporating user feedback and cultural context for enhanced accuracy. This field is crucial for mitigating the spread of harmful content and improving the user experience on online platforms, with ongoing efforts to develop more effective, explainable, and culturally sensitive moderation tools.
Papers
November 7, 2024
November 6, 2024
August 21, 2024
August 10, 2024
June 6, 2024
December 5, 2023
November 14, 2023
July 14, 2023
February 19, 2023
December 27, 2022