Chinese Named Entity Recognition with Character-Level BLSTM and Soft Attention Model

Jize Yin, Senlin Luo, Zhouting Wu, Limin Pan*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

5 Citations (Scopus)

Abstract

Unlike named entity recognition (NER) for English, the absence of word boundaries reduces the final accuracy for Chinese NER. To avoid accumulated error introduced by word segmentation, a deep model extracting character-level features is carefully built and becomes a basis for a new Chinese NER method, which is proposed in this paper. This method converts the raw text to a character vector sequence, extracts global text features with a bidirectional long short-term memory and extracts local text features with a soft attention model. A linear chain conditional random field is also used to label all the characters with the help of the global and local text features. Experiments based on the Microsoft Research Asia (MSRA) dataset are designed and implemented. Results show that the proposed method has good performance compared to other methods, which proves that the global and local text features extracted have a positive influence on Chinese NER. For more variety in the test domains, a resume dataset from Sina Finance is also used to prove the effectiveness of the proposed method.

Original languageEnglish
Pages (from-to)60-71
Number of pages12
JournalJournal of Beijing Institute of Technology (English Edition)
Volume29
Issue number1
DOIs
Publication statusPublished - 1 Mar 2020

Keywords

  • Bidirectional long short-term memory
  • Character-level
  • Chinese
  • Named entity recognition (NER)
  • Soft attention model

Fingerprint

Dive into the research topics of 'Chinese Named Entity Recognition with Character-Level BLSTM and Soft Attention Model'. Together they form a unique fingerprint.

Cite this