Non Toxic
Research on "non-toxic" language focuses on detecting and mitigating harmful content generated by large language models (LLMs), particularly toxic, biased, and offensive language. Current efforts concentrate on developing robust detection models using transformer architectures like BERT and LLMs, exploring methods to reduce toxicity during model training and prompting, and creating comprehensive benchmark datasets reflecting diverse languages and cultural contexts. This research is crucial for ensuring the safe and ethical deployment of LLMs in various applications, mitigating the risks of harmful content generation and promoting responsible AI development.
Papers
January 3, 2024
December 13, 2023
November 29, 2023
November 3, 2023
October 26, 2023
October 20, 2023
September 12, 2023
September 8, 2023
August 16, 2023
July 14, 2023
May 22, 2023
May 11, 2023
April 24, 2023
April 14, 2023
April 11, 2023
March 6, 2023
February 14, 2023
January 30, 2023
December 15, 2022