Towards end-to-end learning for efficient dialogue agent by modeling looking-ahead ability

Zhuoxuan Jiang, Xian Ling Mao*, Ziming Huang, Jie Ma, Shaochun Li

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

3 Citations (Scopus)

Abstract

Learning an efficient manager of dialogue agent from data with little manual intervention is important, especially for goal-oriented dialogues. However, existing methods either take too many manual efforts (e.g. reinforcement learning methods) or cannot guarantee the dialogue efficiency (e.g. sequence-to-sequence methods). In this paper, we address this problem by proposing a novel end-to-end learning model to train a dialogue agent that can look ahead for several future turns and generate an optimal response to make the dialogue efficient. Our method is data-driven and does not require too much manual work for intervention during system design. We evaluate our method on two datasets of different scenarios and the experimental results demonstrate the efficiency of our model.

Original languageEnglish
Title of host publicationSIGDIAL 2019 - 20th Annual Meeting of the Special Interest Group Discourse Dialogue - Proceedings of the Conference
PublisherAssociation for Computational Linguistics (ACL)
Pages133-142
Number of pages10
ISBN (Electronic)9781950737611
Publication statusPublished - 2019
Event20th Annual Meeting of the Special Interest Group on Discourse and Dialogue, SIGDIAL 2019 - Stockholm, Sweden
Duration: 11 Sept 201913 Sept 2019

Publication series

NameSIGDIAL 2019 - 20th Annual Meeting of the Special Interest Group Discourse Dialogue - Proceedings of the Conference

Conference

Conference20th Annual Meeting of the Special Interest Group on Discourse and Dialogue, SIGDIAL 2019
Country/TerritorySweden
CityStockholm
Period11/09/1913/09/19

Fingerprint

Dive into the research topics of 'Towards end-to-end learning for efficient dialogue agent by modeling looking-ahead ability'. Together they form a unique fingerprint.

Cite this