TY - GEN
T1 - Enhancing recurrent neural networks with positional attention for question answering
AU - Chen, Qin
AU - Hu, Qinmin
AU - Huang, Jimmy Xiangji
AU - He, Liang
AU - An, Weijie
N1 - Publisher Copyright:
© 2017 Copyright held by the owner/author(s).
PY - 2017/8/7
Y1 - 2017/8/7
N2 - Attention based recurrent neural networks (RNN) have shown a great success for question answering (QA) in recent years. Although significant improvements have been achieved over the nonattentive models, the position information is not well studied within the attention-based framework. Motivated by the effectiveness of using the word positional context to enhance information retrieval, we assume that if a word in the question (i.e., question word) occurs in an answer sentence, the neighboring words should be given more attention since they intuitively contain more valuable information for question answering than those far away. Based on this assumption, we propose a positional attention based RNN model, which incorporates the positional context of the question words into the answers' attentive representations. Experiments on two benchmark datasets show the great advantages of our proposed model. Specifically, we achieve a maximum improvement of 8.83% over the classical attention based RNN model in terms of mean average precision. Furthermore, our model is comparable to if not better than the state-of-The-Art approaches for question answering.
AB - Attention based recurrent neural networks (RNN) have shown a great success for question answering (QA) in recent years. Although significant improvements have been achieved over the nonattentive models, the position information is not well studied within the attention-based framework. Motivated by the effectiveness of using the word positional context to enhance information retrieval, we assume that if a word in the question (i.e., question word) occurs in an answer sentence, the neighboring words should be given more attention since they intuitively contain more valuable information for question answering than those far away. Based on this assumption, we propose a positional attention based RNN model, which incorporates the positional context of the question words into the answers' attentive representations. Experiments on two benchmark datasets show the great advantages of our proposed model. Specifically, we achieve a maximum improvement of 8.83% over the classical attention based RNN model in terms of mean average precision. Furthermore, our model is comparable to if not better than the state-of-The-Art approaches for question answering.
UR - https://www.scopus.com/pages/publications/85029382274
U2 - 10.1145/3077136.3080699
DO - 10.1145/3077136.3080699
M3 - 会议稿件
AN - SCOPUS:85029382274
T3 - SIGIR 2017 - Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval
SP - 993
EP - 996
BT - SIGIR 2017 - Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval
PB - Association for Computing Machinery, Inc
T2 - 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2017
Y2 - 7 August 2017 through 11 August 2017
ER -