TY - JOUR
T1 - Knowledge enhanced zero-resource machine translation using image-pivoting
AU - Huang, Ping
AU - Zhao, Jing
AU - Sun, Shilinag
AU - Lin, Yichu
N1 - Publisher Copyright:
© 2022, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
PY - 2023/4
Y1 - 2023/4
N2 - Zero resource machine translation usually means that there are no parallel corpora in the training of machine translation models, which can be solved with the help of extra information such as images. However, the ambiguity in the text, together with the irrelevant information in images, may cause the problem of translation errors of some key words. In order to alleviate the problem of image-text alignment deviation caused by word ambiguity, we introduce knowledge entities as an extra modality for the source language to enhance the representations of the source text to clarify its semantics. Specifically, we use additional multi-modal information including images and knowledge entities as an auxiliary hint for the source text in the Transformer-based zero-resource translation framework. We also solve the problem of the structural difference between the training and inference stages to handle the cases where there is no longer visual information in the inference stage. The proposed method achieves state-of-the-art BLEU scores in the field of zero-resource machine translation with the image as the pivot.
AB - Zero resource machine translation usually means that there are no parallel corpora in the training of machine translation models, which can be solved with the help of extra information such as images. However, the ambiguity in the text, together with the irrelevant information in images, may cause the problem of translation errors of some key words. In order to alleviate the problem of image-text alignment deviation caused by word ambiguity, we introduce knowledge entities as an extra modality for the source language to enhance the representations of the source text to clarify its semantics. Specifically, we use additional multi-modal information including images and knowledge entities as an auxiliary hint for the source text in the Transformer-based zero-resource translation framework. We also solve the problem of the structural difference between the training and inference stages to handle the cases where there is no longer visual information in the inference stage. The proposed method achieves state-of-the-art BLEU scores in the field of zero-resource machine translation with the image as the pivot.
KW - Knowledge enhancement
KW - Multi-modal learning
KW - Transformer
KW - Zero-resource machine translation
UR - https://www.scopus.com/pages/publications/85135275754
U2 - 10.1007/s10489-022-03997-0
DO - 10.1007/s10489-022-03997-0
M3 - 文章
AN - SCOPUS:85135275754
SN - 0924-669X
VL - 53
SP - 7484
EP - 7496
JO - Applied Intelligence
JF - Applied Intelligence
IS - 7
ER -