Abstract

Language models have been used in many natural language processing applications. In recent years, the recurrent neural network based language models have defeated the conventional n-gram based techniques. However, it is difficult for neural network architectures to use linguistic annotations. We try to incorporate part-of-speech features in recurrent neural network language model, and use them to predict the next word. Specifically, we proposed a parallel structure which contains two recurrent neural networks, one for word sequence modeling and another for part-of-speech sequence modeling. The state of part-of-speech network helped improve the word sequence's prediction. Experiments show that the proposed method performs better than the traditional recurrent network on perplexity and is better at reranking machine translation outputs.

Original languageEnglish
Pages140-147
Number of pages8
Publication statusPublished - 2019
Event31st Pacific Asia Conference on Language, Information and Computation, PACLIC 2017 - Cebu City, Philippines
Duration: 16 Nov 201718 Nov 2017

Conference

Conference31st Pacific Asia Conference on Language, Information and Computation, PACLIC 2017
Country/TerritoryPhilippines
CityCebu City
Period16/11/1718/11/17

Fingerprint

Dive into the research topics of 'A parallel recurrent neural network for language modeling with POS tags'. Together they form a unique fingerprint.

Cite this