Making Pretrained Language Models Good Long-tailed Learners

Chen Zhang, Lei Ren, Jingang Wang*, Wei Wu, Dawei Song*

*此作品的通讯作者

科研成果: 会议稿件论文同行评审

4 引用 (Scopus)

摘要

Prompt-tuning has shown appealing performance in few-shot classification by virtue of its capability in effectively exploiting pre-trained knowledge. This motivates us to check the hypothesis that prompt-tuning is also a promising choice for long-tailed classification, since the tail classes are intuitively few-shot ones. To achieve this aim, we conduct empirical studies to examine the hypothesis. The results demonstrate that prompt-tuning makes pretrained language models at least good long-tailed learners. For intuitions on why prompt-tuning can achieve good performance in long-tailed classification, we carry out in-depth analyses by progressively bridging the gap between prompt-tuning and commonly used finetuning. The summary is that the classifier structure and parameterization form the key to making good long-tailed learners, in comparison with the less important input structure. Finally, we verify the applicability of our finding to few-shot classification.

源语言英语
3298-3312
页数15
出版状态已出版 - 2022
活动2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 - Abu Dhabi, 阿拉伯联合酋长国
期限: 7 12月 202211 12月 2022

会议

会议2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022
国家/地区阿拉伯联合酋长国
Abu Dhabi
时期7/12/2211/12/22

指纹

探究 'Making Pretrained Language Models Good Long-tailed Learners' 的科研主题。它们共同构成独一无二的指纹。

引用此

Zhang, C., Ren, L., Wang, J., Wu, W., & Song, D. (2022). Making Pretrained Language Models Good Long-tailed Learners. 3298-3312. 论文发表于 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, 阿拉伯联合酋长国.