Improving pretrained cross-lingual language models via self-labeled word alignment

Zewen Chi*, Li Dong, Bo Zheng*, Shaohan Huang, Xian Ling Mao, Heyan Huang, Furu Wei

*此作品的通讯作者

科研成果: 书/报告/会议事项章节会议稿件同行评审

50 引用 (Scopus)

摘要

The cross-lingual language models are typically pretrained with masked language modeling on multilingual text or parallel sentences. In this paper, we introduce denoising word alignment as a new cross-lingual pre-training task. Specifically, the model first self-labels word alignments for parallel sentences. Then we randomly mask tokens in a bitext pair. Given a masked token, the model uses a pointer network to predict the aligned token in the other language. We alternately perform the above two steps in an expectation-maximization manner. Experimental results show that our method improves cross-lingual transferability on various datasets, especially on the token-level tasks, such as question answering, and structured prediction. Moreover, the model can serve as a pretrained word aligner, which achieves reasonably low error rates on the alignment benchmarks. The code and pretrained parameters are available at github.com/CZWin32768/XLM-Align.

源语言英语
主期刊名ACL-IJCNLP 2021 - 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Proceedings of the Conference
出版商Association for Computational Linguistics (ACL)
3418-3430
页数13
ISBN(电子版)9781954085527
出版状态已出版 - 2021
活动Joint Conference of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL-IJCNLP 2021 - Virtual, Online
期限: 1 8月 20216 8月 2021

出版系列

姓名ACL-IJCNLP 2021 - 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Proceedings of the Conference

会议

会议Joint Conference of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL-IJCNLP 2021
Virtual, Online
时期1/08/216/08/21

指纹

探究 'Improving pretrained cross-lingual language models via self-labeled word alignment' 的科研主题。它们共同构成独一无二的指纹。

引用此