TY - JOUR
T1 - Self-attention-based neural networks for refining the overlength product titles
AU - Lin, Yuming
AU - Fu, Yu
AU - Li, You
AU - Cai, Guoyong
AU - Zhou, Aoying
N1 - Publisher Copyright:
© 2021, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
PY - 2021/7
Y1 - 2021/7
N2 - Online sellers often produce redundant and lengthy product textual titles with extra information on e-commerce platforms to attract the attentions of customers. Such overlength product titles become a problem when they are displayed on mobile applications. In this paper, the problem of refining redundant and overlength product titles is studied to generate concise and informative titles. First, the task of refining the long title is transformed into a sequential classification problem by predicting whether a word in original title will remain in finial short title. Then, a self-attention-based neural network is proposed to extract the most informative words from original title to construct the short title. The proposed basic model is also extended with a gated recurrent unit (GRU) neural network and a gating mechanism to improve the position encoding process and learn the weights of encoding features from different directions. Moreover, an algorithm is designed to construct the datasets for redundant product title compression analysis based on the open dataset LESD4EC. Finally, extensive experiments are implemented on the rebuilt datasets to demonstrate the effectiveness and efficiency of the proposed methods. The experimental results show that the proposed methods significantly outperform the state-of-the-art methods based on the precision, recall, F1 value and the mean absolute error, as well as runtime and space cost.
AB - Online sellers often produce redundant and lengthy product textual titles with extra information on e-commerce platforms to attract the attentions of customers. Such overlength product titles become a problem when they are displayed on mobile applications. In this paper, the problem of refining redundant and overlength product titles is studied to generate concise and informative titles. First, the task of refining the long title is transformed into a sequential classification problem by predicting whether a word in original title will remain in finial short title. Then, a self-attention-based neural network is proposed to extract the most informative words from original title to construct the short title. The proposed basic model is also extended with a gated recurrent unit (GRU) neural network and a gating mechanism to improve the position encoding process and learn the weights of encoding features from different directions. Moreover, an algorithm is designed to construct the datasets for redundant product title compression analysis based on the open dataset LESD4EC. Finally, extensive experiments are implemented on the rebuilt datasets to demonstrate the effectiveness and efficiency of the proposed methods. The experimental results show that the proposed methods significantly outperform the state-of-the-art methods based on the precision, recall, F1 value and the mean absolute error, as well as runtime and space cost.
KW - GRU neural network
KW - Gating mechanism
KW - Product title refinement
KW - Self-attention mechanism
UR - https://www.scopus.com/pages/publications/85107317214
U2 - 10.1007/s11042-021-10908-x
DO - 10.1007/s11042-021-10908-x
M3 - 文章
AN - SCOPUS:85107317214
SN - 1380-7501
VL - 80
SP - 28501
EP - 28519
JO - Multimedia Tools and Applications
JF - Multimedia Tools and Applications
IS - 18
ER -