Improving non-autoregressive machine translation via autoregressive training

Shuheng Wang, Shumin Shi*, Heyan Huang, Wei Zhang

*此作品的通讯作者

科研成果: 期刊稿件会议文章同行评审

摘要

In recent years, non-autoregressive machine translation has attracted many researchers’ attentions. Non-autoregressive translation (NAT) achieves faster decoding speed at the cost of translation accuracy compared with autoregressive translation (AT). Since NAT and AT models have similar architecture, a natural idea is to use AT task assisting NAT task. Previous works use curriculum learning or distillation to improve the performance of NAT model. However, they are complex to follow and diffucult to be integrated into some new works. So in this paper, to make it easy, we introduce a multi-task framework to improve the performance of NAT task. Specially, we use a fully shared encoder-decoder network to train NAT task and AT task simultaneously. To evaluate the performance of our model, we conduct experiments on serval benchmask tasks, including WMT14 EN-DE, WMT16 EN-RO and IWSLT14 DE-EN. The experimental results demonstrate that our model achieves improvements but still keeps simple.

源语言英语
文章编号012045
期刊Journal of Physics: Conference Series
2031
1
DOI
出版状态已出版 - 30 9月 2021
活动2021 2nd International Conference on Signal Processing and Computer Science, SPCS 2021 - Qingdao, Virtual, 中国
期限: 20 8月 202122 8月 2021

指纹

探究 'Improving non-autoregressive machine translation via autoregressive training' 的科研主题。它们共同构成独一无二的指纹。

引用此