Filter-GLAT: Filter Glanced Decoder Output for Non-autoregressive Transformer

  • Zichun Wang
  • , Huanran Zheng
  • , Xiaoling Wang*
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Non-autoregressive machine translation model has achieved significantly faster inference speed compared to the autoregressive translation model. However, its translation quality is degraded compared to the autoregressive translation model. Despite numerous advanced methods are proposed to improve the translation quality of the non-autoregressive translation model, achieving the desired trade-off between quality and efficiency is difficult. In this paper, a Filter Glanced Transformer, named Filter-GLAT, is proposed to tackle this problem. It first refines the glance sampling learning strategy, followed by adopting the Filter learning strategy during training, substantially enhancing the translation quality. As for the inference speed, Filter-GLAT generates predictions with only a single decoding pass, maintaining high speed. Moreover, the Filter learning strategy helps the model narrow the gap between training and inference procedures by modifying the training process. Extensive experiments over translation benchmarks (WMT’14 EN-DE and WMT’16 EN-RO) demonstrate that Filter-GLAT almost strikes the best balance between translation quality and speed.

Original languageEnglish
Title of host publicationWeb and Big Data - 8th International Joint Conference, APWeb-WAIM 2024, Proceedings
EditorsWenjie Zhang, Zhengyi Yang, Xiaoyang Wang, Anthony Tung, Zhonglong Zheng, Hongjie Guo
PublisherSpringer Science and Business Media Deutschland GmbH
Pages59-73
Number of pages15
ISBN (Print)9789819772315
DOIs
StatePublished - 2024
Event8th Asia-Pacific Web and Web-Age Information Management Joint International Conference on Web and Big Data, APWeb-WAIM 2024 - Jinhua, China
Duration: 30 Aug 20241 Sep 2024

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume14961 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference8th Asia-Pacific Web and Web-Age Information Management Joint International Conference on Web and Big Data, APWeb-WAIM 2024
Country/TerritoryChina
CityJinhua
Period30/08/241/09/24

Keywords

  • Efficient Inference
  • Learning Strategy
  • Neural Machine Translation
  • Non-autoregressive Generation

Fingerprint

Dive into the research topics of 'Filter-GLAT: Filter Glanced Decoder Output for Non-autoregressive Transformer'. Together they form a unique fingerprint.

Cite this