Stop filtering: Multi-view attribute-enhanced dialogue learning

Yiwei Li, Bin Sun, Shaoxiong Feng, Kan Li*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

There is a growing interest in improving the conversational ability of models by filtering dialogue corpora. Previous filtering strategies rely on a scoring method to assess and discard samples from one perspective, enabling the model to enhance the corresponding dialogue attributes (e.g., consistency). However, the discarded samples may achieve high scores in other perspectives and can also provide regularization effects on the model learning. In this work, we propose a multi-view attribute-enhanced dialogue learning framework that can capture the attribute-related features steadily and comprehensively. Instead of filtering the raw dataset, our framework introduces adapters to learn knowledge from the attribute-related sub-sets after pre-training the model on the full dataset. Considering the variety of dialogue attributes, we further design a multi-view enhancement mechanism, including multi-view selection and inter-view fusion. It groups the high-quality samples from multiple perspectives, respectively, and enhances different attributes of responses with the corresponding sub-sets and adapters, keeping knowledge independent and allowing flexible integration. Empirical results and analysis show that our framework outperforms the state-of-the-art data filtering methods significantly in terms of enhancing dialogue attributes and fusing view-specific knowledge.

Original languageEnglish
Article number110833
JournalKnowledge-Based Systems
Volume277
DOIs
Publication statusPublished - 9 Oct 2023

Keywords

  • Data filtering
  • Dialogue attributes
  • Open-domain dialogue generation

Fingerprint

Dive into the research topics of 'Stop filtering: Multi-view attribute-enhanced dialogue learning'. Together they form a unique fingerprint.

Cite this