Mathematical Reasoning
Mathematical reasoning in large language models (LLMs) is a burgeoning research area focused on evaluating and improving the ability of these models to solve mathematical problems, encompassing both symbolic and numerical reasoning. Current research emphasizes developing more robust benchmarks that assess not only final accuracy but also the reasoning process itself, including error detection and correction, and exploring various training methods such as reinforcement learning from human feedback and instruction tuning to enhance model performance. This field is significant because advancements in mathematical reasoning capabilities in LLMs have broad implications for various applications, including education, scientific discovery, and automated problem-solving.
Papers
Evaluating LLMs' Mathematical and Coding Competency through Ontology-guided Interventions
Pengfei Hong, Navonil Majumder, Deepanway Ghosal, Somak Aditya, Rada Mihalcea, Soujanya Poria
Augmenting Math Word Problems via Iterative Question Composing
Haoxiong Liu, Yifan Zhang, Yifan Luo, Andrew Chi-Chih Yao
TinyGSM: achieving >80% on GSM8k with small language models
Bingbin Liu, Sebastien Bubeck, Ronen Eldan, Janardhan Kulkarni, Yuanzhi Li, Anh Nguyen, Rachel Ward, Yi Zhang
Modeling Complex Mathematical Reasoning via Large Language Model based MathAgent
Haoran Liao, Qinyi Du, Shaohua Hu, Hao He, Yanyan Xu, Jidong Tian, Yaohui Jin
Fewer is More: Boosting LLM Reasoning with Reinforced Context Pruning
Xijie Huang, Li Lyna Zhang, Kwang-Ting Cheng, Fan Yang, Mao Yang
math-PVS: A Large Language Model Framework to Map Scientific Publications to PVS Theories
Hassen Saidi, Susmit Jha, Tuhin Sahai
SkyMath: Technical Report
Liu Yang, Haihua Yang, Wenjun Cheng, Lei Lin, Chenxia Li, Yifu Chen, Lunan Liu, Jianfei Pan, Tianwen Wei, Biye Li, Liang Zhao, Lijie Wang, Bo Zhu, Guoliang Li, Xuejie Wu, Xilin Luo, Rui Hu