LLM Agent
LLM agents are autonomous systems that combine large language models with the ability to interact with their environment, achieving complex tasks through planning, tool use, and iterative refinement. Current research focuses on improving their reliability and safety, including aligning their behavior with human values, enhancing their decision-making processes (e.g., through Q-value models and tree search algorithms), and mitigating vulnerabilities to adversarial attacks. This field is significant because it bridges the gap between theoretical AI and practical applications, impacting diverse areas such as game development, software testing, healthcare, and financial markets by automating tasks and improving decision-making.
Papers
Prioritizing Safeguarding Over Autonomy: Risks of LLM Agents for Science
Xiangru Tang, Qiao Jin, Kunlun Zhu, Tongxin Yuan, Yichi Zhang, Wangchunshu Zhou, Meng Qu, Yilun Zhao, Jian Tang, Zhuosheng Zhang, Arman Cohan, Zhiyong Lu, Mark Gerstein
LLM Agents can Autonomously Hack Websites
Richard Fang, Rohan Bindu, Akul Gupta, Qiusi Zhan, Daniel Kang
On Evaluating the Integration of Reasoning and Action in LLM Agents with Database Question Answering
Linyong Nan, Ellen Zhang, Weijin Zou, Yilun Zhao, Wenfei Zhou, Arman Cohan
Tailoring with Targeted Precision: Edit-Based Agents for Open-Domain Procedure Customization
Yash Kumar Lal, Li Zhang, Faeze Brahman, Bodhisattwa Prasad Majumder, Peter Clark, Niket Tandon