Language Feedback
Language feedback, the use of natural language to guide and improve large language model (LLM) performance, is a rapidly developing area of research aiming to align LLMs more closely with human intentions and values. Current work focuses on developing methods that leverage language feedback for reward modeling in reinforcement learning, enabling LLMs to self-evaluate and refine their outputs, and improving interactive learning through feedback loops. This research is significant because it offers a more nuanced and efficient approach to LLM training than traditional methods, potentially leading to more reliable and human-aligned AI systems across various applications.
Papers
Adapting LLM Agents with Universal Feedback in Communication
Kuan Wang, Yadong Lu, Michael Santacroce, Yeyun Gong, Chao Zhang, Yelong Shen
SELF: Self-Evolution with Language Feedback
Jianqiao Lu, Wanjun Zhong, Wenyong Huang, Yufei Wang, Qi Zhu, Fei Mi, Baojun Wang, Weichao Wang, Xingshan Zeng, Lifeng Shang, Xin Jiang, Qun Liu