@inproceedings{cecf316fb9334a3399aa0ac86dcd3a30,
title = "Abstractive Summarization Model with Adaptive Sparsemax",
abstract = "Abstractive summarization models mostly rely on Sequence-to-Sequence architectures, in which the softmax function is widely used to transform the model output to simplex. However, softmax{\textquoteright}s output probability distribution often has the long-tail effect especially when the vocabulary size is large. Many unrelated tokens occupy too many probabilities so they will reduce the training efficiency and effect. More recently, some work has begun to design mapping functions to gain sparse output probabilities to ignore these irrelevant tokens. In this paper, we propose Adaptive Sparsemax which can self-adaptively control the sparsity of the model{\textquoteright}s output. Our method combines sparsemax and temperature mechanism, and the temperature value can be learned by the neural network. One of the advantages of our method is that it doesn{\textquoteright}t need any hyperparameter. The experimental result on CNN-Daily Mail and LCSTS dataset shows that our method has better performance on the abstractive summarization task than baseline models.",
keywords = "Abstractive summarization, Adaptive sparsemax, Seq2Seq",
author = "Shiqi Guo and Yumeng Si and Jing Zhao",
note = "Publisher Copyright: {\textcopyright} 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.; 11th CCF International Conference on Natural Language Processing and Chinese Computing, NLPCC 2022 ; Conference date: 24-09-2022 Through 25-09-2022",
year = "2022",
doi = "10.1007/978-3-031-17120-8\_62",
language = "英语",
isbn = "9783031171192",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer Science and Business Media Deutschland GmbH",
pages = "810--821",
editor = "Wei Lu and Shujian Huang and Yu Hong and Xiabing Zhou",
booktitle = "Natural Language Processing and Chinese Computing - 11th CCF International Conference, NLPCC 2022, Proceedings",
address = "德国",
}