Can Monolingual Pretrained Models Help Cross-Lingual Classification?

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

6 Citations (Scopus)

Abstract

Multilingual pretrained language models (such as multilingual BERT) have achieved impressive results for cross-lingual transfer. However, due to the constant model capacity, multilingual pre-training usually lags behind the monolingual competitors. In this work, we present two approaches to improve zero-shot cross-lingual classification, by transferring the knowledge from monolingual pretrained models to multilingual ones. Experimental results on two cross-lingual classification benchmarks show that our methods outperform vanilla multilingual fine-tuning.

Original languageEnglish
Title of host publicationProceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, AACL-IJCNLP 2020
EditorsKam-Fai Wong, Kevin Knight, Hua Wu
PublisherAssociation for Computational Linguistics (ACL)
Pages12-17
Number of pages6
ISBN (Electronic)9781952148910
DOIs
Publication statusPublished - 2020
Event1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, AACL-IJCNLP 2020 - Virtual, Online, China
Duration: 4 Dec 20207 Dec 2020

Publication series

NameProceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, AACL-IJCNLP 2020

Conference

Conference1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, AACL-IJCNLP 2020
Country/TerritoryChina
CityVirtual, Online
Period4/12/207/12/20

Fingerprint

Dive into the research topics of 'Can Monolingual Pretrained Models Help Cross-Lingual Classification?'. Together they form a unique fingerprint.

Cite this