Self Generated Response
Self-generated response research focuses on enabling language models (LLMs) to improve their outputs without extensive human supervision, primarily by leveraging the models' own evaluations of their generated text. Current approaches explore methods like self-play, where LLMs refine their responses by comparing self-generated alternatives, and self-alignment techniques that use internal knowledge for evaluating and improving factuality or adherence to specified principles. This research aims to reduce reliance on costly human feedback datasets and improve the overall quality and reliability of LLM outputs, impacting both the efficiency of LLM training and the trustworthiness of their applications.
Papers
May 23, 2024
April 22, 2024
April 4, 2024
February 14, 2024
January 2, 2024
October 10, 2023
November 22, 2022