摘要
Recent advances in distilling pretrained language models have discovered that, besides the expressiveness of knowledge, the student-friendliness should be taken into consideration to realize a truly knowledgeable teacher. Based on a pilot study, we find that over-parameterized teachers can produce expressive yet student-unfriendly knowledge and are thus limited in overall knowledgeableness. To remove the parameters that result in student-unfriendliness, we propose a sparse teacher trick under the guidance of an overall knowledgeable score for each teacher parameter. The knowledgeable score is essentially an interpolation of the expressiveness and student-friendliness scores. The aim is to ensure that the expressive parameters are retained while the student-unfriendly ones are removed. Extensive experiments on the GLUE benchmark show that the proposed sparse teachers can be dense with knowledge and lead to students with compelling performance in comparison with a series of competitive baselines.
源语言 | 英语 |
---|---|
页 | 3904-3915 |
页数 | 12 |
出版状态 | 已出版 - 2022 |
活动 | 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 - Abu Dhabi, 阿拉伯联合酋长国 期限: 7 12月 2022 → 11 12月 2022 |
会议
会议 | 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 |
---|---|
国家/地区 | 阿拉伯联合酋长国 |
市 | Abu Dhabi |
时期 | 7/12/22 → 11/12/22 |