Enhancing Small Model Performance in Educational Classification Tasks through Knowledge Distillation

  • Haoxin Xu
  • , Changyong Qi
  • , Bingqian Jiang
  • , Tong Liu
  • , Longwei Zheng*
  • , Xiaoqing Gu
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

As the demand for precision, efficiency, and low-cost solutions in educational classification tasks continues to grow, enhancing model performance has become a critical focus of research. While large language models excel in these tasks, their high cost and resource requirements limit widespread application. This study proposes a Knowledge-Enhanced Distillation (KED) method, utilizing ChatGPT-4, ChatGPT-4o, and Llama3 as teacher models, and three different sizes of BERT models as student models. The method was validated across three real-world educational datasets. The results demonstrate that the KED method significantly improves the accuracy and F1 scores of small models in educational text classification tasks, while also substantially reducing computational costs and resource consumption. Notably, the KED method shows exceptional performance in scenarios involving few-shot learning and class imbalance. The innovation of this study lies in applying the KED method to educational classification tasks, filling a gap in current research and highlighting its significant potential for practical application in educational contexts.

Original languageEnglish
Title of host publication2025 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2025 - Proceedings
EditorsBhaskar D Rao, Isabel Trancoso, Gaurav Sharma, Neelesh B. Mehta
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798350368741
DOIs
StatePublished - 2025
Event2025 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2025 - Hyderabad, India
Duration: 6 Apr 202511 Apr 2025

Publication series

NameICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
ISSN (Print)1520-6149

Conference

Conference2025 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2025
Country/TerritoryIndia
CityHyderabad
Period6/04/2511/04/25

Keywords

  • Educational Application
  • Few-Shot Learning
  • Knowledge Distillation
  • Large Language Model

Fingerprint

Dive into the research topics of 'Enhancing Small Model Performance in Educational Classification Tasks through Knowledge Distillation'. Together they form a unique fingerprint.

Cite this