Textual Feedback
Textual feedback is being explored as a powerful mechanism for improving the performance and alignment of various machine learning models, particularly large language models (LLMs) and image-text alignment models. Current research focuses on developing algorithms that effectively incorporate textual feedback into training and inference, leveraging techniques like reinforcement learning and fine-tuning with both binary and richer, descriptive feedback. This work is significant because it allows for more efficient and effective model training and adaptation to user preferences, leading to improvements in tasks ranging from code generation and summarization to image retrieval and toxicity reduction.
Papers
October 4, 2024
July 24, 2024
December 5, 2023
May 17, 2023
March 28, 2023
November 14, 2022
April 29, 2022