TY - GEN
T1 - Routing Based Context Selection for Document-Level Neural Machine Translation
AU - Fei, Weilun
AU - Jian, Ping
AU - Zhu, Xiaoguang
AU - Lin, Yi
N1 - Publisher Copyright:
© 2021, Springer Nature Singapore Pte Ltd.
PY - 2021
Y1 - 2021
N2 - Most of the existing methods of document-level neural machine translation (NMT) integrate more textual information by extending the scope of sentence encoding. Usually, the sentence-level representation is incorporated (via attention or gate mechanism) in these methods, which makes them straightforward but rough, and it is difficult to distinguish useful contextual information from noises. Furthermore, the longer the encoding length is, the more difficult it is for the model to grasp the inter-dependency between sentences. In this paper, a document-level NMT method based on a routing algorithm is presented, which can automatically select context information. The routing mechanism endows the current source sentence with the ability to decide which words can become its context. This leads the method to merge the inter-sentence dependencies in a more flexible and elegant way, and model local structure information more effectively. At the same time, this structured information selection mechanism will also alleviate the possible problems caused by long-distance encoding. Experimental results show that our method is 2.91 BLEU higher than the Transformer model on the public dataset of ZH-EN, and is superior to most of the state-of-the-art document-level NMT models.
AB - Most of the existing methods of document-level neural machine translation (NMT) integrate more textual information by extending the scope of sentence encoding. Usually, the sentence-level representation is incorporated (via attention or gate mechanism) in these methods, which makes them straightforward but rough, and it is difficult to distinguish useful contextual information from noises. Furthermore, the longer the encoding length is, the more difficult it is for the model to grasp the inter-dependency between sentences. In this paper, a document-level NMT method based on a routing algorithm is presented, which can automatically select context information. The routing mechanism endows the current source sentence with the ability to decide which words can become its context. This leads the method to merge the inter-sentence dependencies in a more flexible and elegant way, and model local structure information more effectively. At the same time, this structured information selection mechanism will also alleviate the possible problems caused by long-distance encoding. Experimental results show that our method is 2.91 BLEU higher than the Transformer model on the public dataset of ZH-EN, and is superior to most of the state-of-the-art document-level NMT models.
KW - Document-Level Neural Machine Translation
KW - Natural Language Processing
KW - Routing Algorithm
UR - http://www.scopus.com/inward/record.url?scp=85119343559&partnerID=8YFLogxK
U2 - 10.1007/978-981-16-7512-6_7
DO - 10.1007/978-981-16-7512-6_7
M3 - Conference contribution
AN - SCOPUS:85119343559
SN - 9789811675119
T3 - Communications in Computer and Information Science
SP - 77
EP - 91
BT - Machine Translation - 17th China Conference, CCMT 2021, Revised Selected Papers
A2 - Su, Jinsong
A2 - Sennrich, Rico
PB - Springer Science and Business Media Deutschland GmbH
T2 - 17th China Conference on Machine Translation, CCMT 2021
Y2 - 8 October 2021 through 10 October 2021
ER -