Non Toxic
Research on "non-toxic" language focuses on detecting and mitigating harmful content generated by large language models (LLMs), particularly toxic, biased, and offensive language. Current efforts concentrate on developing robust detection models using transformer architectures like BERT and LLMs, exploring methods to reduce toxicity during model training and prompting, and creating comprehensive benchmark datasets reflecting diverse languages and cultural contexts. This research is crucial for ensuring the safe and ethical deployment of LLMs in various applications, mitigating the risks of harmful content generation and promoting responsible AI development.
Papers
October 29, 2024
October 16, 2024
October 8, 2024
October 1, 2024
September 25, 2024
September 21, 2024
September 15, 2024
August 1, 2024
July 22, 2024
July 8, 2024
May 29, 2024
May 17, 2024
May 16, 2024
April 29, 2024
April 22, 2024
April 9, 2024
March 25, 2024
February 12, 2024
February 10, 2024
January 23, 2024