TY - GEN
T1 - Dual-Attention Graph Convolutional Network
AU - Zhang, Xueya
AU - Zhang, Tong
AU - Zhao, Wenting
AU - Cui, Zhen
AU - Yang, Jian
N1 - Publisher Copyright:
© 2020, Springer Nature Switzerland AG.
PY - 2020
Y1 - 2020
N2 - Graph convolutional networks (GCNs) have shown the powerful ability in text structure representation and effectively facilitate the task of text classification. However, challenges still exist in adapting GCN on learning discriminative features from texts due to the main issue of graph variants incurred by the textual complexity and diversity. In this paper, we propose a dual-attention GCN to model the structural information of various texts as well as tackle the graph-invariant problem through embedding two types of attention mechanisms, i.e. the connection-attention and hop-attention, into the classic GCN. To encode various connection patterns between neighbour words, connection-attention adaptively imposes different weights specified to neighbourhoods of each word, which captures the short-term dependencies. On the other hand, the hop-attention applies scaled coefficients to different scopes during the graph diffusion process to make the model learn more about the distribution of context, which captures long-term semantics in an adaptive way. Extensive experiments are conducted on five widely used datasets to evaluate our dual-attention GCN, and the achieved state-of-the-art performance verifies the effectiveness of dual-attention mechanisms.
AB - Graph convolutional networks (GCNs) have shown the powerful ability in text structure representation and effectively facilitate the task of text classification. However, challenges still exist in adapting GCN on learning discriminative features from texts due to the main issue of graph variants incurred by the textual complexity and diversity. In this paper, we propose a dual-attention GCN to model the structural information of various texts as well as tackle the graph-invariant problem through embedding two types of attention mechanisms, i.e. the connection-attention and hop-attention, into the classic GCN. To encode various connection patterns between neighbour words, connection-attention adaptively imposes different weights specified to neighbourhoods of each word, which captures the short-term dependencies. On the other hand, the hop-attention applies scaled coefficients to different scopes during the graph diffusion process to make the model learn more about the distribution of context, which captures long-term semantics in an adaptive way. Extensive experiments are conducted on five widely used datasets to evaluate our dual-attention GCN, and the achieved state-of-the-art performance verifies the effectiveness of dual-attention mechanisms.
KW - Dual-attention
KW - Graph convolutional networks
KW - Text classification
UR - http://www.scopus.com/inward/record.url?scp=85081550040&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-41299-9_19
DO - 10.1007/978-3-030-41299-9_19
M3 - Conference contribution
AN - SCOPUS:85081550040
SN - 9783030412982
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 238
EP - 251
BT - Pattern Recognition - 5th Asian Conference, ACPR 2019, Revised Selected Papers
A2 - Palaiahnakote, Shivakumara
A2 - Sanniti di Baja, Gabriella
A2 - Wang, Liang
A2 - Yan, Wei Qi
PB - Springer
T2 - 5th Asian Conference on Pattern Recognition, ACPR 2019
Y2 - 26 November 2019 through 29 November 2019
ER -