Mathematical Reasoning
Mathematical reasoning in large language models (LLMs) is a burgeoning research area focused on evaluating and improving the ability of these models to solve mathematical problems, encompassing both symbolic and numerical reasoning. Current research emphasizes developing more robust benchmarks that assess not only final accuracy but also the reasoning process itself, including error detection and correction, and exploring various training methods such as reinforcement learning from human feedback and instruction tuning to enhance model performance. This field is significant because advancements in mathematical reasoning capabilities in LLMs have broad implications for various applications, including education, scientific discovery, and automated problem-solving.
Papers
Novice Learner and Expert Tutor: Evaluating Math Reasoning Abilities of Large Language Models with Misconceptions
Naiming Liu, Shashank Sonkar, Zichao Wang, Simon Woodhead, Richard G. Baraniuk
MathVista: Evaluating Mathematical Reasoning of Foundation Models in Visual Contexts
Pan Lu, Hritik Bansal, Tony Xia, Jiacheng Liu, Chunyuan Li, Hannaneh Hajishirzi, Hao Cheng, Kai-Wei Chang, Michel Galley, Jianfeng Gao
A Mechanistic Interpretation of Arithmetic Reasoning in Language Models using Causal Mediation Analysis
Alessandro Stolfo, Yonatan Belinkov, Mrinmaya Sachan
Calc-X and Calcformers: Empowering Arithmetical Chain-of-Thought through Interaction with Symbolic Systems
Marek Kadlčík, Michal Štefánik, Ondřej Sotolář, Vlastimil Martinek