Diversifying Neural Dialogue Generation via Negative Distillation

Yiwei Li, Shaoxiong Feng, Bin Sun, Kan Li

科研成果: 书/报告/会议事项章节会议稿件同行评审

8 引用 (Scopus)

摘要

Generative dialogue models suffer badly from the generic response problem, limiting their applications to a few toy scenarios. Recently, an interesting approach, namely negative training, has been proposed to alleviate this problem by reminding the model not to generate high-frequency responses during training. However, its performance is hindered by two issues, ignoring low-frequency but generic responses and bringing low-frequency but meaningless responses. In this paper, we propose a novel negative training paradigm, called negative distillation, to keep the model away from the undesirable generic responses while avoiding the above problems. First, we introduce a negative teacher model that can produce query-wise generic responses, and then the student model is required to maximize the distance with multi-level negative knowledge. Empirical results show that our method outperforms previous negative training methods significantly.

源语言英语
主期刊名NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics
主期刊副标题Human Language Technologies, Proceedings of the Conference
出版商Association for Computational Linguistics (ACL)
407-418
页数12
ISBN(电子版)9781955917711
出版状态已出版 - 2022
活动2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2022 - Seattle, 美国
期限: 10 7月 202215 7月 2022

出版系列

姓名NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference

会议

会议2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2022
国家/地区美国
Seattle
时期10/07/2215/07/22

指纹

探究 'Diversifying Neural Dialogue Generation via Negative Distillation' 的科研主题。它们共同构成独一无二的指纹。

引用此