Alleviating repetitive tokens in non-autoregressive machine translation with unlikelihood training

Shuheng Wang, Shumin Shi*, Heyan Huang

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

摘要

In recent years, significant progress has been made in the field of non-autoregressive machine translations. However, the accuracy of non-autoregressive models still lags behind their autoregressive counterparts. This discrepancy can be attributed to the abundance of repetitive tokens in the target sequences generated by non-autoregressive models. In this study, we delve into this phenomenon and propose a novel approach to train a non-autoregressive model using unlikelihood loss. We evaluate our method on three widely used benchmark tasks. The experimental results demonstrating that our proposed approach significantly reduces the number of repetitive tokens while improving the overall performance of non-autoregressive machine translations. Compared to the baseline model ”Mask-Predict”, the average number of repetitions on IWSLT 14 DE→EN valid set is reduced from 0.48 to 0.17, resulting in a remarkable 62% decrease.

源语言英语
页(从-至)4681-4688
页数8
期刊Soft Computing
28
5
DOI
出版状态已出版 - 3月 2024

指纹

探究 'Alleviating repetitive tokens in non-autoregressive machine translation with unlikelihood training' 的科研主题。它们共同构成独一无二的指纹。

引用此

Wang, S., Shi, S., & Huang, H. (2024). Alleviating repetitive tokens in non-autoregressive machine translation with unlikelihood training. Soft Computing, 28(5), 4681-4688. https://doi.org/10.1007/s00500-023-09490-1