@inproceedings{c11714fa848b487d901cf1915e1b804f,
title = "Making Pretrained Language Models Good Long-tailed Learners",
abstract = "Prompt-tuning has shown appealing performance in few-shot classification by virtue of its capability in effectively exploiting pre-trained knowledge. This motivates us to check the hypothesis that prompt-tuning is also a promising choice for long-tailed classification, since the tail classes are intuitively few-shot ones. To achieve this aim, we conduct empirical studies to examine the hypothesis. The results demonstrate that prompt-tuning makes pretrained language models at least good long-tailed learners. For intuitions on why prompt-tuning can achieve good performance in long-tailed classification, we carry out in-depth analyses by progressively bridging the gap between prompt-tuning and commonly used finetuning. The summary is that the classifier structure and parameterization form the key to making good long-tailed learners, in comparison with the less important input structure. Finally, we verify the applicability of our finding to few-shot classification.",
author = "Chen Zhang and Lei Ren and Jingang Wang and Wei Wu and Dawei Song",
note = "Publisher Copyright: {\textcopyright} 2022 Association for Computational Linguistics.; 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 ; Conference date: 07-12-2022 Through 11-12-2022",
year = "2022",
doi = "10.18653/v1/2022.emnlp-main.217",
language = "English",
series = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022",
publisher = "Association for Computational Linguistics (ACL)",
pages = "3298--3312",
editor = "Yoav Goldberg and Zornitsa Kozareva and Yue Zhang",
booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022",
address = "United States",
}