Hierarchical text-label integrated attention network for document classification

Changjin Gong, Kaize Shi, Zhendong Niu

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

7 Citations (Scopus)

Abstract

Recurrent neural networks (RNN) and convolutional neural networks (CNN) have been extensively used on text classification to capture the local and long-range dependencies. Recent work has demonstrated the superiority of self-attention networks (SAN) owing to their highly parallelizable computation and excellent performance. However, SAN has difficulty capturing meaningful semantic relationships over very long sequences, and the memory requirement grows rapidly in line with the sequence length. To solve these limitations of SAN in processing long document sequence, this paper proposes four novel ideas and build a hierarchical text-label integrated attention network(HLAN). Firstly, a hierarchical architecture is introduced to map the hierarchy of document, which effectively shortens the sequence length of each process. Secondly, the attention weights are calculated in the joint embedding space of text and label. Thirdly, a multi-head soft attention is proposed to compress the sequence encoded by self-attention into a single vector. Finally, a loss term called class loss is given and combined with cross entropy loss. HLAN achieves competitive results over the compared strong baseline methods on 4 out of 5 benchmark datasets, which verifies the effectiveness of HLAN for document classification, in terms of both accuracy and memory requirement.

Original languageEnglish
Title of host publicationHPCCT 2019 - 3rd High Performance Computing and Cluster Technologies Conference and BDAI 2019 - 2nd International Conference on Big Data and Artificial Intelligence
PublisherAssociation for Computing Machinery
Pages254-260
Number of pages7
ISBN (Electronic)9781450371858
DOIs
Publication statusPublished - 22 Jun 2019
Event3rd High Performance Computing and Cluster Technologies Conference, HPCCT 2019 and the 2nd International Conference on Big Data and Artificial Intelligence, BDAI 2019 - Guangzhou, China
Duration: 22 Jun 201924 Jun 2019

Publication series

NameACM International Conference Proceeding Series

Conference

Conference3rd High Performance Computing and Cluster Technologies Conference, HPCCT 2019 and the 2nd International Conference on Big Data and Artificial Intelligence, BDAI 2019
Country/TerritoryChina
CityGuangzhou
Period22/06/1924/06/19

Keywords

  • Class loss
  • Document classification
  • Hierarchical
  • Memory requirement
  • Self-attention networks
  • Textlabel integrated

Fingerprint

Dive into the research topics of 'Hierarchical text-label integrated attention network for document classification'. Together they form a unique fingerprint.

Cite this