Reasoning Ability
Reasoning ability in large language models (LLMs) is a burgeoning research area focused on evaluating and enhancing the capacity of these models to perform multi-step inferences and solve complex problems requiring logical deduction and inductive learning. Current research emphasizes benchmarking LLMs on diverse tasks, including mathematical reasoning, commonsense reasoning, and following procedures, often employing techniques like chain-of-thought prompting and knowledge distillation to improve performance. Understanding and improving LLM reasoning is crucial for building more reliable and trustworthy AI systems with broader applications across various fields, from scientific discovery to decision-making support.
Papers
Learning to Reason Iteratively and Parallelly for Complex Visual Reasoning Scenarios
Shantanu Jaiswal, Debaditya Roy, Basura Fernando, Cheston Tan
Disentangling Memory and Reasoning Ability in Large Language Models
Mingyu Jin, Weidi Luo, Sitao Cheng, Xinyi Wang, Wenyue Hua, Ruixiang Tang, William Yang Wang, Yongfeng Zhang
Patience Is The Key to Large Language Model Reasoning
Yijiong Yu
HumanEval-V: Evaluating Visual Understanding and Reasoning Abilities of Large Multimodal Models Through Coding Tasks
Fengji Zhang, Linquan Wu, Huiyu Bai, Guancheng Lin, Xiao Li, Xiao Yu, Yue Wang, Bei Chen, Jacky Keung
Reversal of Thought: Enhancing Large Language Models with Preference-Guided Reverse Reasoning Warm-up
Jiahao Yuan, Dehui Du, Hao Zhang, Zixiang Di, Usman Naseem
Which Programming Language and What Features at Pre-training Stage Affect Downstream Logical Inference Performance?
Fumiya Uchiyama, Takeshi Kojima, Andrew Gambardella, Qi Cao, Yusuke Iwasawa, Yutaka Matsuo
Do great minds think alike? Investigating Human-AI Complementarity in Question Answering with CAIMIRA
Maharshi Gor, Hal Daumé III, Tianyi Zhou, Jordan Boyd-Graber
Enhance Reasoning by Learning from Mistakes: Peer-Review Knowledge Distillation from Multiple Large Language Models
Zhuochun Li, Yuelyu Ji, Rui Meng, Daqing He
ProcBench: Benchmark for Multi-Step Reasoning and Following Procedure
Ippei Fujisawa, Sensho Nobe, Hiroki Seto, Rina Onda, Yoshiaki Uchida, Hiroki Ikoma, Pei-Chun Chien, Ryota Kanai