Complex Reasoning Task
Complex reasoning tasks challenge large language models (LLMs) to perform multi-step inferences and solve problems requiring the integration of diverse knowledge and logical operations. Current research focuses on improving LLMs' reasoning abilities through techniques like chain-of-thought prompting, reinforcement learning with refined credit assignment, and the integration of symbolic reasoning methods with neural networks. These advancements aim to enhance the reliability and generalizability of LLMs for applications ranging from scientific discovery and medical diagnosis to automated problem-solving and decision-making, ultimately contributing to a deeper understanding of artificial intelligence and its potential societal impact.
Papers
PEARL: Prompting Large Language Models to Plan and Execute Actions Over Long Documents
Simeng Sun, Yang Liu, Shuohang Wang, Chenguang Zhu, Mohit Iyyer
ChatCoT: Tool-Augmented Chain-of-Thought Reasoning on Chat-based Large Language Models
Zhipeng Chen, Kun Zhou, Beichen Zhang, Zheng Gong, Wayne Xin Zhao, Ji-Rong Wen