Botfip-LLM: An enhanced multimodal scientific computing framework leveraging knowledge distillation from large language models

  • Tianhao Chen
  • , Pengbo Xu*
  • , Haibiao Zheng
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

In recent years, the introduction of AI technologies has brought transformative changes to scientific computing. However, AI models typically focus on single-task and single-modal data processing, limiting their application. To address this, multimodal scientific computing frameworks have become a trend. The Botfip framework aligns function images with symbolic operation trees through multimodal training, extracting deep scientific information. However, Botfip struggles with processing Formula Strings, leading to inadequate understanding in multimodal learning. To enhance Botfip's learning of Formula Strings and expand its applicability to related tasks, we propose the Botfip-LLM framework based on knowledge distillation, incorporating pre-trained large language models for aligning operation tree data. Experimental analysis shows that the choice of LLM is crucial, with ChatGLM-2 outperforming others in training and testing. Botfip-LLM not only improves performance, generalization, and extrapolation over the original Botfip model but also significantly enhances applicability to Formula String-related tasks, enabling more diverse task handling.

Original languageEnglish
Article number114012
JournalKnowledge-Based Systems
Volume326
DOIs
StatePublished - 27 Sep 2025

Keywords

  • Knowledge distillation
  • Large language models
  • Multimodal learning
  • Scientific computing

Fingerprint

Dive into the research topics of 'Botfip-LLM: An enhanced multimodal scientific computing framework leveraging knowledge distillation from large language models'. Together they form a unique fingerprint.

Cite this