TY - JOUR
T1 - Botfip-LLM
T2 - An enhanced multimodal scientific computing framework leveraging knowledge distillation from large language models
AU - Chen, Tianhao
AU - Xu, Pengbo
AU - Zheng, Haibiao
N1 - Publisher Copyright:
© 2025 Elsevier B.V.
PY - 2025/9/27
Y1 - 2025/9/27
N2 - In recent years, the introduction of AI technologies has brought transformative changes to scientific computing. However, AI models typically focus on single-task and single-modal data processing, limiting their application. To address this, multimodal scientific computing frameworks have become a trend. The Botfip framework aligns function images with symbolic operation trees through multimodal training, extracting deep scientific information. However, Botfip struggles with processing Formula Strings, leading to inadequate understanding in multimodal learning. To enhance Botfip's learning of Formula Strings and expand its applicability to related tasks, we propose the Botfip-LLM framework based on knowledge distillation, incorporating pre-trained large language models for aligning operation tree data. Experimental analysis shows that the choice of LLM is crucial, with ChatGLM-2 outperforming others in training and testing. Botfip-LLM not only improves performance, generalization, and extrapolation over the original Botfip model but also significantly enhances applicability to Formula String-related tasks, enabling more diverse task handling.
AB - In recent years, the introduction of AI technologies has brought transformative changes to scientific computing. However, AI models typically focus on single-task and single-modal data processing, limiting their application. To address this, multimodal scientific computing frameworks have become a trend. The Botfip framework aligns function images with symbolic operation trees through multimodal training, extracting deep scientific information. However, Botfip struggles with processing Formula Strings, leading to inadequate understanding in multimodal learning. To enhance Botfip's learning of Formula Strings and expand its applicability to related tasks, we propose the Botfip-LLM framework based on knowledge distillation, incorporating pre-trained large language models for aligning operation tree data. Experimental analysis shows that the choice of LLM is crucial, with ChatGLM-2 outperforming others in training and testing. Botfip-LLM not only improves performance, generalization, and extrapolation over the original Botfip model but also significantly enhances applicability to Formula String-related tasks, enabling more diverse task handling.
KW - Knowledge distillation
KW - Large language models
KW - Multimodal learning
KW - Scientific computing
UR - https://www.scopus.com/pages/publications/105010190986
U2 - 10.1016/j.knosys.2025.114012
DO - 10.1016/j.knosys.2025.114012
M3 - 文章
AN - SCOPUS:105010190986
SN - 0950-7051
VL - 326
JO - Knowledge-Based Systems
JF - Knowledge-Based Systems
M1 - 114012
ER -