Abstract
Neural Machine Translation (NMT) brings promising improvements in translation quality, but until recently, these models rely on large-scale parallel corpora. As such corpora only exist on a handful of language pairs, the translation performance is far from the desired effect in the majority of low-resource languages. Thus, developing low-resource language translation techniques is crucial and it has become a popular research field in neural machine translation. In this article, we make an overall review of existing deep learning techniques in low-resource NMT. We first show the research status as well as some widely used low-resource datasets. Then, we categorize the existing methods and show some representative works detailedly. Finally, we summarize the common characters among them and outline the future directions in this field.
Original language | English |
---|---|
Article number | 103 |
Journal | ACM Transactions on Asian and Low-Resource Language Information Processing |
Volume | 21 |
Issue number | 5 |
DOIs | |
Publication status | Published - 15 Nov 2022 |
Keywords
- Low-resource
- data augmentation
- neural machine translation
- pivot-based methods
- semi-supervised
- transfer learning
- unsupervised