Toxicity Detection Datasets
Toxicity detection datasets are crucial for training and evaluating models that identify harmful online content, aiming to mitigate the spread of hate speech, misinformation, and other forms of toxic language across various modalities (text, audio). Current research focuses on creating more comprehensive and nuanced datasets, addressing limitations like language diversity, contextual understanding, and the influence of annotator bias, often employing transformer-based models and graph neural networks for improved performance. These advancements are vital for enhancing the safety and reliability of online platforms and AI systems, impacting content moderation, brand safety, and the development of more responsible AI technologies.
Papers
November 12, 2024
October 2, 2024
June 27, 2024
February 23, 2024
February 10, 2024
January 10, 2024
November 22, 2023
October 26, 2023
March 27, 2023
June 27, 2022
May 24, 2022
February 8, 2022
November 19, 2021
November 15, 2021