TY - GEN
T1 - Modeling Zero-Shot Relation Classification as a Multiple-Choice Problem
AU - Lan, Yuquan
AU - Li, Dongxu
AU - Zhang, Yunqi
AU - Zhao, Hui
AU - Zhao, Gang
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - Zero-shot relation classification (ZeroRC) aims to infer the semantic relations between entity pairs in sentences, while the relation sets at the training and testing stages are disjoint. It is a crucial task in information extraction and is much more challenging than traditional relation classification. We propose a novel method named MC-BERT to model ZeroRC as a Multiple-Choice problem employing BERT as the backbone model. A semantic template is designed to infuse the information of entities and context. It serves as the question stem of the multiple-choice, followed by a relation label which serves as a choice. Moreover, we propose a grouping strategy to improve training efficiency. We perform comprehensive experiments on two datasets, Wiki-ZSL and FewRel. The results show that our proposed method significantly outperforms previous works. Specifically, it achieves performance gains ranging from 1.17% to 6.87% in the F1 score with only 40% of the model parameters against to state-of-the-art method, demonstrating the simplicity and effectiveness of our proposed method.
AB - Zero-shot relation classification (ZeroRC) aims to infer the semantic relations between entity pairs in sentences, while the relation sets at the training and testing stages are disjoint. It is a crucial task in information extraction and is much more challenging than traditional relation classification. We propose a novel method named MC-BERT to model ZeroRC as a Multiple-Choice problem employing BERT as the backbone model. A semantic template is designed to infuse the information of entities and context. It serves as the question stem of the multiple-choice, followed by a relation label which serves as a choice. Moreover, we propose a grouping strategy to improve training efficiency. We perform comprehensive experiments on two datasets, Wiki-ZSL and FewRel. The results show that our proposed method significantly outperforms previous works. Specifically, it achieves performance gains ranging from 1.17% to 6.87% in the F1 score with only 40% of the model parameters against to state-of-the-art method, demonstrating the simplicity and effectiveness of our proposed method.
KW - information extraction
KW - natural language processing
KW - relation classification
KW - zero-shot learning
UR - https://www.scopus.com/pages/publications/85169599325
U2 - 10.1109/IJCNN54540.2023.10191459
DO - 10.1109/IJCNN54540.2023.10191459
M3 - 会议稿件
AN - SCOPUS:85169599325
T3 - Proceedings of the International Joint Conference on Neural Networks
BT - IJCNN 2023 - International Joint Conference on Neural Networks, Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2023 International Joint Conference on Neural Networks, IJCNN 2023
Y2 - 18 June 2023 through 23 June 2023
ER -