基于注意力机制的概念化句嵌入研究

Translated title of the contribution: Conceptual Sentence Embeddings Based on Attention Mechanism

Ya Shen Wang, He Yan Huang*, Chong Feng, Qiang Zhou

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

5 Citations (Scopus)

Abstract

Most sentence embedding models typically represent each sentence only using word surface, which makes these models indiscriminative for ubiquitous homonymy and polysemy. In order to enhance representation capability of sentence, we employ short-text conceptualization algorithm to assign associated concepts for each sentence in the text corpus, and then learn conceptual sentence embedding (CSE). Hence, this semantic representation is more expressive than some widely-used text representation models such as latent topic model, especially for short-text. Moreover, we further extend CSE models by utilizing an attention mechanism that select relevant words within the context to make more efficient prediction. In the experiments, we evaluate the CSE models on three tasks, text classification and information retrieval. The experimental results show that the proposed models outperform typical sentence embed-ding models.

Translated title of the contributionConceptual Sentence Embeddings Based on Attention Mechanism
Original languageChinese (Traditional)
Pages (from-to)1390-1400
Number of pages11
JournalZidonghua Xuebao/Acta Automatica Sinica
Volume46
Issue number7
DOIs
Publication statusPublished - 1 Jul 2020

Fingerprint

Dive into the research topics of 'Conceptual Sentence Embeddings Based on Attention Mechanism'. Together they form a unique fingerprint.

Cite this