Parallel Corpora Alignment Framework for Multilingual and Robust Automatic Dialogue Evaluation

Xinglin Wang, Jiayi Shi, Peiwen Yuan, Kan Li*

*Corresponding author for this work

Research output: Contribution to conferencePaperpeer-review

Abstract

Open-domain automatic dialogue evaluation plays an important role in dialogue systems. While recent efforts are being put into making learning-based evaluation metrics correlate better with human evaluation, robust metrics for parallel corpora and multiple domains remain unexplored. Parallel corpora refer to corpora that express the same idea in different ways (e.g., translation, paraphrasing and back-translation). In this paper, we propose Parallel Corpora Alignment Framework (PCAF), which improves the consistency and robustness of model evaluation on parallel corpora. Firstly, parallel corpora are aligned in semantic space through parallel-corpora-aligned contrastive learning. Then, parallel-corpora-aligned distillation on multiple datasets is applied to further improve model’s generalization ability across multiple data domains. Our approach ranks second on the final test data of DSTC11 track4 sub-task1 ("Multilingual Automatic Evaluation Metrics", turn-level) and third on the sub-task2 ("Robust Automatic Evaluation Metrics", turn-level), which proves the strong generalization ability and robustness of our proposed approach.

Original languageEnglish
Pages123-132
Number of pages10
Publication statusPublished - 2023
Event11th Dialog System Technology Challenge, DSTC 2023 - Prague, Czech Republic
Duration: 11 Sept 2023 → …

Conference

Conference11th Dialog System Technology Challenge, DSTC 2023
Country/TerritoryCzech Republic
CityPrague
Period11/09/23 → …

Fingerprint

Dive into the research topics of 'Parallel Corpora Alignment Framework for Multilingual and Robust Automatic Dialogue Evaluation'. Together they form a unique fingerprint.

Cite this