External Feedback
External feedback, crucial for guiding machine learning models towards desired behaviors, is a central focus of current research. Studies explore various feedback modalities, including comparative preferences, step-level explanations, and even crowdsourced large language model (LLM) assessments, employing techniques like Proximal Policy Optimization (PPO) and novel algorithms designed for relative feedback. This research aims to improve model performance, efficiency, and alignment with human values, impacting fields ranging from robotics and reinforcement learning to natural language processing and human-computer interaction. The ultimate goal is to create more robust, adaptable, and human-centered AI systems.
Papers
September 20, 2024
July 20, 2024
July 11, 2024
June 14, 2024
May 26, 2024
May 23, 2024
May 1, 2024
April 22, 2024
January 4, 2024
December 11, 2023
October 3, 2023
September 30, 2023
June 26, 2023
May 19, 2023
April 4, 2023
August 21, 2022