OnMKD: An Online Mutual Knowledge Distillation Framework for Passage Retrieval

Jiali Deng, Dongyang Li, Taolin Zhang, Xiaofeng He

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Dense passage retriever recalls a set of relevant passages from a large corpus according to a natural language question. The dual-encoder architecture is prevalent in dense passage retrievers, which is based on large-scale pre-trained language models (PLMs). However, existing PLMs usually have thick structures and bulky parameters, resulting in large memory and time consumption. To overcome the limitation of PLMs, in this paper we apply online distillation to passage retrieval and propose an Online Mutual Knowledge Distillation framework (OnMKD). Specifically, we obtain a lightweight retriever by simultaneously updating two peer networks with the same dual-encoder structure and different initial parameters, named Online Mutual Knowledge Refinement. To further interact with the latent knowledge of intermediate layers, we utilize a novel cross-wise contrastive loss to alternate the representation of questions and passages. Experimental results indicate that our framework outperforms other small baselines with the same number of layers on multiple QA benchmarks. Compared to the heavy PLMs, OnMKD significantly accelerates the inference process and reduces storage requirements with only a slight sacrifice in performance.

Original languageEnglish
Title of host publicationNatural Language Processing and Chinese Computing - 12th National CCF Conference, NLPCC 2023, Proceedings
EditorsFei Liu, Nan Duan, Qingting Xu, Yu Hong
PublisherSpringer Science and Business Media Deutschland GmbH
Pages719-731
Number of pages13
ISBN (Print)9783031446955
DOIs
StatePublished - 2023
Event12th National CCF Conference on Natural Language Processing and Chinese Computing, NLPCC 2023 - Foshan, China
Duration: 12 Oct 202315 Oct 2023

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume14303 LNAI
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference12th National CCF Conference on Natural Language Processing and Chinese Computing, NLPCC 2023
Country/TerritoryChina
CityFoshan
Period12/10/2315/10/23

Keywords

  • Cross-wise contrastive learning
  • Mutual learning
  • Online knowledge distillation
  • Passage retrieval

Fingerprint

Dive into the research topics of 'OnMKD: An Online Mutual Knowledge Distillation Framework for Passage Retrieval'. Together they form a unique fingerprint.

Cite this