XLM-E: Cross-lingual Language Model Pre-training via ELECTRA

Zewen Chi, Shaohan Huang, Li Dong, Shuming Ma, Bo Zheng, Saksham Singhal, Payal Bajaj, Xia Song, Xian Ling Mao, Heyan Huang, Furu Wei

科研成果: 书/报告/会议事项章节会议稿件同行评审

50 引用 (Scopus)

摘要

In this paper, we introduce ELECTRA-style tasks (Clark et al., 2020b) to cross-lingual language model pre-training. Specifically, we present two pre-training tasks, namely multilingual replaced token detection, and translation replaced token detection. Besides, we pretrain the model, named as XLM-E, on both multilingual and parallel corpora. Our model outperforms the baseline models on various cross-lingual understanding tasks with much less computation cost. Moreover, analysis shows that XLM-E tends to obtain better cross-lingual transferability.

源语言英语
主期刊名ACL 2022 - 60th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers)
编辑Smaranda Muresan, Preslav Nakov, Aline Villavicencio
出版商Association for Computational Linguistics (ACL)
6170-6182
页数13
ISBN(电子版)9781955917216
出版状态已出版 - 2022
活动60th Annual Meeting of the Association for Computational Linguistics, ACL 2022 - Dublin, 爱尔兰
期限: 22 5月 202227 5月 2022

出版系列

姓名Proceedings of the Annual Meeting of the Association for Computational Linguistics
1
ISSN(印刷版)0736-587X

会议

会议60th Annual Meeting of the Association for Computational Linguistics, ACL 2022
国家/地区爱尔兰
Dublin
时期22/05/2227/05/22

指纹

探究 'XLM-E: Cross-lingual Language Model Pre-training via ELECTRA' 的科研主题。它们共同构成独一无二的指纹。

引用此