Metacognitive symbolic distillation framework for multi-choice machine reading comprehension

Jiacheng Yao, Xin Xu, Guoxiu He

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

Symbolic knowledge distillation can transfer the reasoning abilities of large language models (LLMs) effectively to smaller models. However, in the context of multi-choice machine reading comprehension (MMRC), traditional distillation methods focus primarily on learning from the rationales of the correct options generated by the large teacher model, overlooking the educational significance of reasoning behind incorrect options. In human education, metacognition emphasizes the importance of actively identifying errors, enhancing the overall understanding. Inspired by this approach, we propose an innovative framework that incorporates metacognition into symbolic distillation. Initially, we prompt the teacher LLM to generate rationales for all options in the MMRC dataset. Subsequently, the small student model is fine-tuned using these rationales, including those for incorrect options. Our experiments on two MMRC datasets demonstrate that this framework improves the performance of the small student model significantly compared to standard fine-tuned and distilled models. We further find that when the student model is sufficiently large, upgrading the teacher model could yield further improvements. However, the effectiveness of our framework is constrained by the performance of the teacher model on more complex MMRC tasks.

Original languageEnglish
Article number113130
JournalKnowledge-Based Systems
Volume312
DOIs
StatePublished - 15 Mar 2025

Keywords

  • Knowledge distillation
  • Large language model
  • Metacognition
  • Multi-choice machine reading comprehension
  • Symbolic distillation

Fingerprint

Dive into the research topics of 'Metacognitive symbolic distillation framework for multi-choice machine reading comprehension'. Together they form a unique fingerprint.

Cite this