TT-GCN: Temporal-Tightly Graph Convolutional Network for Emotion Recognition From Gaits

Tong Zhang, Yelin Chen, Shuzhen Li, Xiping Hu, C. L.Philip Chen*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

5 Citations (Scopus)

Abstract

The human gait reflects substantial information about individual emotions. Current gait emotion recognition methods focus on capturing gait topology information and ignore the importance of fine-grained temporal features. This article proposes the temporal-tightly graph convolutional network (TT-GCN) to extract temporal features. TT-GCN comprises three significant mechanisms: the causal temporal convolution network (casual-TCN), the walking direction recognition auxiliary task, and the feature mapping layer. To obtain tight temporal dependencies and enhance the relevance among gait periods, the causal-TCN is introduced. Based on the assumption of emotional consistency in the walking directions, the auxiliary task is proposed to enhance the ability of fine-grained feature extraction. Through the feature mapping layer, affective features can be mapped into the appropriate representation and fused with deep learning features. TT-GCN shows the best performance across five comprehensive metrics. All experimental results verify the necessity and feasibility of exploring fine-grained temporal feature extraction.

Original languageEnglish
Pages (from-to)4300-4314
Number of pages15
JournalIEEE Transactions on Computational Social Systems
Volume11
Issue number3
DOIs
Publication statusPublished - 1 Jun 2024

Keywords

  • Emotion recognition
  • gait
  • graph convolutional network (GCN)

Fingerprint

Dive into the research topics of 'TT-GCN: Temporal-Tightly Graph Convolutional Network for Emotion Recognition From Gaits'. Together they form a unique fingerprint.

Cite this