Leveraging pattern associations for word embedding models

Qian Liu, Heyan Huang, Yang Gao*, Xiaochi Wei, Ruiying Geng

*Corresponding author for this work

Research output: Contribution to journalConference articlepeer-review

2 Citations (Scopus)

Abstract

Word embedding method has been shown powerful to capture words association, and facilitated numerous applications by effectively bridging lexical gaps. Word semantic is encoded with vectors and modeled based on n-gram language models, as a result it only takes into consideration of words co-occurrences in a shallow slide windows. However, the assumption of the language modelling ignores valuable associations between words in a long distance beyond n-gram coverage. In this paper, we argue that it is beneficial to jointly modeling both surrounding context and flexible associative patterns so that the model can cover long distance and intensive association. We propose a novel approach to combine associated patterns for word embedding method via joint training objection. We apply our model for query expansion in document retrieval task. Experimental results show that the proposed method can perform significantly better than the state-of-the-arts baseline models.

Original languageEnglish
Pages (from-to)423-438
Number of pages16
JournalLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume10177 LNCS
DOIs
Publication statusPublished - 2017
Event22nd International Conference on Database Systems for Advanced Applications, DASFAA 2017 - Suzhou, China
Duration: 27 Mar 201730 Mar 2017

Fingerprint

Dive into the research topics of 'Leveraging pattern associations for word embedding models'. Together they form a unique fingerprint.

Cite this