TY - GEN
T1 - Connective Prediction for Implicit Discourse Relation Recognition via Knowledge Distillation
AU - Wu, Hongyi
AU - Zhou, Hao
AU - Lan, Man
AU - Wu, Yuanbin
AU - Zhang, Yadong
N1 - Publisher Copyright:
© 2023 Association for Computational Linguistics.
PY - 2023
Y1 - 2023
N2 - Implicit discourse relation recognition (IDRR) remains a challenging task in discourse analysis due to the absence of connectives. Most existing methods utilize one-hot labels as the sole optimization target, ignoring the internal association among connectives. Besides, these approaches spend lots of effort on template construction, negatively affecting the generalization capability. To address these problems, we propose a novel Connective Prediction via Knowledge Distillation (CP-KD) approach to instruct large-scale pre-trained language models (PLMs) mining the latent correlations between connectives and discourse relations, which is meaningful for IDRR. Experimental results on the PDTB 2.0/3.0 and CoNLL 2016 datasets show that our method significantly outperforms the state-of-the-art models on coarse-grained and fine-grained discourse relations. Moreover, our approach can be transferred to explicit discourse relation recognition (EDRR) and achieve acceptable performance. Our code is released in https://github.com/cubenlp/CP_KD-for-IDRR.
AB - Implicit discourse relation recognition (IDRR) remains a challenging task in discourse analysis due to the absence of connectives. Most existing methods utilize one-hot labels as the sole optimization target, ignoring the internal association among connectives. Besides, these approaches spend lots of effort on template construction, negatively affecting the generalization capability. To address these problems, we propose a novel Connective Prediction via Knowledge Distillation (CP-KD) approach to instruct large-scale pre-trained language models (PLMs) mining the latent correlations between connectives and discourse relations, which is meaningful for IDRR. Experimental results on the PDTB 2.0/3.0 and CoNLL 2016 datasets show that our method significantly outperforms the state-of-the-art models on coarse-grained and fine-grained discourse relations. Moreover, our approach can be transferred to explicit discourse relation recognition (EDRR) and achieve acceptable performance. Our code is released in https://github.com/cubenlp/CP_KD-for-IDRR.
UR - https://www.scopus.com/pages/publications/85174410172
U2 - 10.18653/v1/2023.acl-long.325
DO - 10.18653/v1/2023.acl-long.325
M3 - 会议稿件
AN - SCOPUS:85174410172
T3 - Proceedings of the Annual Meeting of the Association for Computational Linguistics
SP - 5908
EP - 5923
BT - Long Papers
PB - Association for Computational Linguistics (ACL)
T2 - 61st Annual Meeting of the Association for Computational Linguistics, ACL 2023
Y2 - 9 July 2023 through 14 July 2023
ER -