TY - JOUR
T1 - Mathematical Language Models
T2 - A Survey
AU - Liu, Wentao
AU - Hu, Hanglei
AU - Zhou, Jie
AU - Ding, Yuyang
AU - Li, Junsong
AU - Zeng, Jiayi
AU - He, Mengliang
AU - Chen, Qin
AU - Jiang, Bo
AU - Zhou, Aimin
AU - He, Liang
N1 - Publisher Copyright:
© 2025 Copyright held by the owner/author(s). Publication rights licensed to ACM.
PY - 2025/12/8
Y1 - 2025/12/8
N2 - In recent years, there has been remarkable progress in leveraging Language Models (LMs), encompassing Pre-trained Language Models (PLMs) and Large-scale Language Models (LLMs), within the domain of mathematics. This article conducts a comprehensive survey of mathematical LMs, systematically categorizing pivotal research endeavors from two distinct perspectives: tasks and methodologies. The landscape reveals a large number of proposed mathematical LLMs, which are further delineated into instruction learning, tool-based methods, fundamental CoT techniques, advanced CoT methodologies, and multi-modal methods. To comprehend the benefits of mathematical LMs more thoroughly, we carry out an in-depth contrast of their characteristics and performance. In addition, our survey entails the compilation of over 60 mathematical datasets, including training datasets, benchmark datasets, and augmented datasets. Addressing the primary challenges and delineating future trajectories within the field of mathematical LMs, this survey is poised to facilitate and inspire future innovation among researchers invested in advancing this domain.
AB - In recent years, there has been remarkable progress in leveraging Language Models (LMs), encompassing Pre-trained Language Models (PLMs) and Large-scale Language Models (LLMs), within the domain of mathematics. This article conducts a comprehensive survey of mathematical LMs, systematically categorizing pivotal research endeavors from two distinct perspectives: tasks and methodologies. The landscape reveals a large number of proposed mathematical LLMs, which are further delineated into instruction learning, tool-based methods, fundamental CoT techniques, advanced CoT methodologies, and multi-modal methods. To comprehend the benefits of mathematical LMs more thoroughly, we carry out an in-depth contrast of their characteristics and performance. In addition, our survey entails the compilation of over 60 mathematical datasets, including training datasets, benchmark datasets, and augmented datasets. Addressing the primary challenges and delineating future trajectories within the field of mathematical LMs, this survey is poised to facilitate and inspire future innovation among researchers invested in advancing this domain.
KW - language models
KW - LLMs
KW - Mathematics
KW - pre-trained
KW - survey
UR - https://www.scopus.com/pages/publications/105026654651
U2 - 10.1145/3773985
DO - 10.1145/3773985
M3 - 文献综述
AN - SCOPUS:105026654651
SN - 0360-0300
VL - 58
JO - ACM Computing Surveys
JF - ACM Computing Surveys
IS - 6
M1 - 146
ER -