Mathematical Language Models: A Survey

  • Wentao Liu*
  • , Hanglei Hu
  • , Jie Zhou
  • , Yuyang Ding
  • , Junsong Li
  • , Jiayi Zeng
  • , Mengliang He
  • , Qin Chen
  • , Bo Jiang
  • , Aimin Zhou
  • , Liang He
  • *Corresponding author for this work

Research output: Contribution to journalReview articlepeer-review

Abstract

In recent years, there has been remarkable progress in leveraging Language Models (LMs), encompassing Pre-trained Language Models (PLMs) and Large-scale Language Models (LLMs), within the domain of mathematics. This article conducts a comprehensive survey of mathematical LMs, systematically categorizing pivotal research endeavors from two distinct perspectives: tasks and methodologies. The landscape reveals a large number of proposed mathematical LLMs, which are further delineated into instruction learning, tool-based methods, fundamental CoT techniques, advanced CoT methodologies, and multi-modal methods. To comprehend the benefits of mathematical LMs more thoroughly, we carry out an in-depth contrast of their characteristics and performance. In addition, our survey entails the compilation of over 60 mathematical datasets, including training datasets, benchmark datasets, and augmented datasets. Addressing the primary challenges and delineating future trajectories within the field of mathematical LMs, this survey is poised to facilitate and inspire future innovation among researchers invested in advancing this domain.

Original languageEnglish
Article number146
JournalACM Computing Surveys
Volume58
Issue number6
DOIs
StatePublished - 8 Dec 2025

Keywords

  • language models
  • LLMs
  • Mathematics
  • pre-trained
  • survey

Fingerprint

Dive into the research topics of 'Mathematical Language Models: A Survey'. Together they form a unique fingerprint.

Cite this