TY - GEN
T1 - ParaSum
T2 - Knowledge Science, Engineering and Management - 16th International Conference, KSEM 2023, Proceedings
AU - Tang, Moming
AU - Wang, Chengyu
AU - Wang, Jianing
AU - Chen, Cen
AU - Gao, Ming
AU - Qian, Weining
N1 - Publisher Copyright:
© 2023, The Author(s), under exclusive license to Springer Nature Switzerland AG.
PY - 2023
Y1 - 2023
N2 - Existing extractive summarization methods achieve state-of-the-art (SOTA) performance with pre-trained language models (PLMs) and sufficient training data. However, PLM-based methods are known to be data-hungry and often fail to deliver satisfactory results in low-resource scenarios. Constructing a high-quality summarization dataset with human-authored reference summaries is a prohibitively expensive task. To address these challenges, this paper proposes a novel paradigm for low-resource extractive summarization, called ParaSum. This paradigm reformulates text summarization as textual paraphrasing, aligning the text summarization task with the self-supervised Next Sentence Prediction (NSP) task of PLMs. This approach minimizes the training gap between the summarization model and PLMs, enabling a more effective probing of the knowledge encoded within PLMs and enhancing the summarization performance. Furthermore, to relax the requirement for large amounts of training data, we introduce a simple yet efficient model and align the training paradigm of summarization to textual paraphrasing to facilitate network-based transfer learning. Extensive experiments over two widely used benchmarks (i.e., CNN/DailyMail, Xsum) and a recent open-sourced high-quality Chinese benchmark (i.e., CNewSum) show that ParaSum consistently outperforms existing PLM-based summarization methods in all low-resource settings, demonstrating its effectiveness over different types of datasets.
AB - Existing extractive summarization methods achieve state-of-the-art (SOTA) performance with pre-trained language models (PLMs) and sufficient training data. However, PLM-based methods are known to be data-hungry and often fail to deliver satisfactory results in low-resource scenarios. Constructing a high-quality summarization dataset with human-authored reference summaries is a prohibitively expensive task. To address these challenges, this paper proposes a novel paradigm for low-resource extractive summarization, called ParaSum. This paradigm reformulates text summarization as textual paraphrasing, aligning the text summarization task with the self-supervised Next Sentence Prediction (NSP) task of PLMs. This approach minimizes the training gap between the summarization model and PLMs, enabling a more effective probing of the knowledge encoded within PLMs and enhancing the summarization performance. Furthermore, to relax the requirement for large amounts of training data, we introduce a simple yet efficient model and align the training paradigm of summarization to textual paraphrasing to facilitate network-based transfer learning. Extensive experiments over two widely used benchmarks (i.e., CNN/DailyMail, Xsum) and a recent open-sourced high-quality Chinese benchmark (i.e., CNewSum) show that ParaSum consistently outperforms existing PLM-based summarization methods in all low-resource settings, demonstrating its effectiveness over different types of datasets.
KW - extractive summarization
KW - low-resource scenarios
KW - pre-trained language model
KW - textual paraphrasing
KW - transfer learning
UR - https://www.scopus.com/pages/publications/85171159033
U2 - 10.1007/978-3-031-40289-0_9
DO - 10.1007/978-3-031-40289-0_9
M3 - 会议稿件
AN - SCOPUS:85171159033
SN - 9783031402883
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 106
EP - 119
BT - Knowledge Science, Engineering and Management - 16th International Conference, KSEM 2023, Proceedings
A2 - Jin, Zhi
A2 - Jiang, Yuncheng
A2 - Ma, Wenjun
A2 - Buchmann, Robert Andrei
A2 - Ghiran, Ana-Maria
A2 - Bi, Yaxin
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 16 August 2023 through 18 August 2023
ER -