LLM Agent
LLM agents are autonomous systems that combine large language models with the ability to interact with their environment, achieving complex tasks through planning, tool use, and iterative refinement. Current research focuses on improving their reliability and safety, including aligning their behavior with human values, enhancing their decision-making processes (e.g., through Q-value models and tree search algorithms), and mitigating vulnerabilities to adversarial attacks. This field is significant because it bridges the gap between theoretical AI and practical applications, impacting diverse areas such as game development, software testing, healthcare, and financial markets by automating tasks and improving decision-making.
Papers
Towards General Computer Control: A Multimodal Agent for Red Dead Redemption II as a Case Study
Weihao Tan, Ziluo Ding, Wentao Zhang, Boyu Li, Bohan Zhou, Junpeng Yue, Haochong Xia, Jiechuan Jiang, Longtao Zheng, Xinrun Xu, Yifei Bi, Pengjie Gu, Xinrun Wang, Börje F. Karlsson, Bo An, Zongqing Lu
OPEx: A Component-Wise Analysis of LLM-Centric Agents in Embodied Instruction Following
Haochen Shi, Zhiyuan Sun, Xingdi Yuan, Marc-Alexandre Côté, Bang Liu
ChatCite: LLM Agent with Human Workflow Guidance for Comparative Literature Summary
Yutong Li, Lu Chen, Aiwei Liu, Kai Yu, Lijie Wen