Neural machine translation with Gumbel Tree-LSTM based encoder

Research output: Contribution to journalArticlepeer-review

31 Citations (Scopus)

Abstract

Neural machine translation has improved the translation accuracy greatly and received great attention of the machine translation community. Tree-based translation models aim to model the syntactic or semantic relation among long-distance words or phrases in a sentence. However, it faces the difficulties of expensive manual annotation cost and poor automatic annotation accuracy. In this paper, we focus on how to encode a source sentence into a vector in a unsupervised-tree way and then decode it into a target sentence. Our model incorporates Gumbel Tree-LSTM, which can learn how to compose tree structures from plain text without any tree annotation. We evaluate the proposed model on both spoken and news corpora, and show that the performance of our proposed model outperforms the attentional seq2seq model and the Transformer base model.

Original languageEnglish
Article number102811
JournalJournal of Visual Communication and Image Representation
Volume71
DOIs
Publication statusPublished - Aug 2020

Keywords

  • Gumbel Tree-LSTM
  • Neural machine translation
  • Tree to sequence

Fingerprint

Dive into the research topics of 'Neural machine translation with Gumbel Tree-LSTM based encoder'. Together they form a unique fingerprint.

Cite this