Automatic Evaluate Dialogue Appropriateness by Using Dialogue Act

Bao Chen, Yuanjie Wang, Zeming Liu, Yuhang Guo*

*此作品的通讯作者

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

Evaluation of dialogue systems requires assessing various aspects, among which appropriateness holds significance as a core element of communicative language competence. However, current evaluations heavily rely on human judgments, which are time-consuming, labor-intensive, prone to biases, and lacking objectivity. In this paper, we introduce Dialogue Act Appropriateness (DAA), a novel method that utilizes the underlying patterns of dialogue act transitions to evaluate the appropriateness of chatbot responses. We learn transition patterns from human-human dialogue corpora, evaluating chatbot appropriateness by measuring the similarity of their transition patterns to those observed in human-human dialogues. To validate DAA, we annotate a test dataset by manually evaluating the appropriateness of dialogues from multiple chatbot systems. The experimental results demonstrate a strong correlation between our evaluation metric and human ratings, establishing the reliability of DAA as a measure of dialogue appropriateness.

源语言英语
主期刊名Findings of the Association for Computational Linguistics
主期刊副标题EMNLP 2023
出版商Association for Computational Linguistics (ACL)
7361-7372
页数12
ISBN(电子版)9798891760615
DOI
出版状态已出版 - 2023
活动2023 Findings of the Association for Computational Linguistics: EMNLP 2023 - Singapore, 新加坡
期限: 6 12月 202310 12月 2023

出版系列

姓名Findings of the Association for Computational Linguistics: EMNLP 2023

会议

会议2023 Findings of the Association for Computational Linguistics: EMNLP 2023
国家/地区新加坡
Singapore
时期6/12/2310/12/23

指纹

探究 'Automatic Evaluate Dialogue Appropriateness by Using Dialogue Act' 的科研主题。它们共同构成独一无二的指纹。

引用此