LLM Policy
LLM policy research focuses on developing methods to control and improve the behavior of large language models, ensuring safe, helpful, and unbiased outputs. Current research explores techniques like iterative policy design inspired by mapmaking, reinforcement learning approaches that prioritize diverse solutions, and methods for filtering self-generated training data to enhance model performance. These advancements are crucial for mitigating risks associated with LLMs and enabling their responsible deployment across various applications, from code generation and legal advice to healthcare and autonomous systems.
Papers
November 1, 2024
October 29, 2024
October 15, 2024
September 26, 2024
September 5, 2024
August 19, 2024
August 17, 2024
August 12, 2024
July 23, 2024
July 22, 2024
July 17, 2024
July 3, 2024
June 9, 2024
April 14, 2024
March 26, 2024
March 11, 2024
February 2, 2024
December 30, 2023
October 5, 2023