Generation with Dynamic Vocabulary

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

We introduce a new dynamic vocabulary for language models.It can involve arbitrary text spans during generation.These text spans act as basic generation bricks, akin to tokens in the traditional static vocabularies.We show that, the ability to generate multi-tokens atomically improve both generation quality and efficiency (compared to the standard language model, the MAUVE metric is increased by 25%, the latency is decreased by 20%).The dynamic vocabulary can be deployed in a plug-and-play way, thus is attractive for various downstream applications.For example, we demonstrate that dynamic vocabulary can be applied to different domains in a training-free manner.It also helps to generate reliable citations in question answering tasks (substantially enhancing citation results without compromising answer accuracy).

Original languageEnglish
Title of host publicationEMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference
EditorsYaser Al-Onaizan, Mohit Bansal, Yun-Nung Chen
PublisherAssociation for Computational Linguistics (ACL)
Pages18931-18948
Number of pages18
ISBN (Electronic)9798891761643
DOIs
StatePublished - 2024
Event2024 Conference on Empirical Methods in Natural Language Processing, EMNLP 2024 - Hybrid, Miami, United States
Duration: 12 Nov 202416 Nov 2024

Publication series

NameEMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference

Conference

Conference2024 Conference on Empirical Methods in Natural Language Processing, EMNLP 2024
Country/TerritoryUnited States
CityHybrid, Miami
Period12/11/2416/11/24

Fingerprint

Dive into the research topics of 'Generation with Dynamic Vocabulary'. Together they form a unique fingerprint.

Cite this