Non Negative Textual Response
Non-negative textual response research focuses on generating accurate, helpful, and unbiased text outputs from large language models (LLMs), addressing issues like hallucinations (fabricating information) and biases. Current research emphasizes improving the faithfulness of LLM responses to input context, often using retrieval-augmented generation (RAG) or fine-tuning techniques to enhance model accuracy and reduce reliance on inherent biases. This work is crucial for building trustworthy LLMs applicable to various fields, including healthcare, education, and customer service, where reliable and unbiased information is paramount.
Papers
June 9, 2023
June 2, 2023
May 28, 2023
May 26, 2023
May 21, 2023
May 12, 2023
May 6, 2023
May 4, 2023
April 29, 2023
March 15, 2023
February 22, 2023
November 26, 2022
October 31, 2022
October 10, 2022
September 29, 2022
September 7, 2022
August 22, 2022
August 4, 2022
August 1, 2022
June 10, 2022