TY - JOUR
T1 - Two-Channel Feature Extraction Convolutional Neural Network for Facial Expression Recognition
AU - Liu, Chang
AU - Hirota, Kaoru
AU - Wang, Bo
AU - Dai, Yaping
AU - Jia, Zhiyang
N1 - Publisher Copyright:
© 2020 Fuji Technology Press. All rights reserved.
PY - 2020/11
Y1 - 2020/11
N2 - An emotion recognition framework based on a two-channel convolutional neural network (CNN) is proposed to detect the affective state of humans through facial expressions. The framework consists of three parts, i.e., the frontal face detection module, the feature extraction module, and the classification module. The feature extraction module contains two channels: one is for raw face images and the other is for texture feature images. The local binary pattern (LBP) images are utilized for texture feature extraction to enrich facial features and improve the network performance. The attention mechanism is adopted in both CNN feature extraction channels to highlight the features that are related to facial expressions. Moreover, arcface loss function is integrated into the proposed network to increase the inter-class distance and decrease the inner-class distance of facial features. The experiments conducted on the two public databases, FER2013 and CK+, demonstrate that the proposed method outperforms the previous methods, with the accuracies of 72.56% and 94.24%, respectively. The improvement in emotion recognition accuracy makes our approach applicable to service robots.
AB - An emotion recognition framework based on a two-channel convolutional neural network (CNN) is proposed to detect the affective state of humans through facial expressions. The framework consists of three parts, i.e., the frontal face detection module, the feature extraction module, and the classification module. The feature extraction module contains two channels: one is for raw face images and the other is for texture feature images. The local binary pattern (LBP) images are utilized for texture feature extraction to enrich facial features and improve the network performance. The attention mechanism is adopted in both CNN feature extraction channels to highlight the features that are related to facial expressions. Moreover, arcface loss function is integrated into the proposed network to increase the inter-class distance and decrease the inner-class distance of facial features. The experiments conducted on the two public databases, FER2013 and CK+, demonstrate that the proposed method outperforms the previous methods, with the accuracies of 72.56% and 94.24%, respectively. The improvement in emotion recognition accuracy makes our approach applicable to service robots.
KW - Convolutional neural network
KW - Facial expression recognition
KW - Local binary pattern
KW - Texture feature
UR - http://www.scopus.com/inward/record.url?scp=85097572156&partnerID=8YFLogxK
U2 - 10.20965/JACIII.2020.P0792
DO - 10.20965/JACIII.2020.P0792
M3 - Article
AN - SCOPUS:85097572156
SN - 1343-0130
VL - 24
SP - 792
EP - 801
JO - Journal of Advanced Computational Intelligence and Intelligent Informatics
JF - Journal of Advanced Computational Intelligence and Intelligent Informatics
IS - 6
ER -