TY - JOUR
T1 - MPDRM
T2 - A Multi-Scale Personalized Depression Recognition Model via facial movements
AU - Liu, Zhenyu
AU - Chen, Bailin
AU - Zhang, Shimao
AU - Yuan, Jiaqian
AU - Wu, Yang
AU - Cai, Hanshu
AU - Chen, Xin
AU - Liu, Lin
AU - Zhao, Yimiao
AU - Mei, Huan
AU - Deng, Jiahui
AU - Bao, Yanping
AU - Hu, Bin
N1 - Publisher Copyright:
© 2025
PY - 2025/6/1
Y1 - 2025/6/1
N2 - Automatic depression recognition based on facial movements in videos has become a research hotspot. However, existing methods tend to confuse individual inherent facial behavioral habits with characteristics specific to depression, which leads to misjudgments. To address this, we propose a Multi-scale Personalized Depression Recognition Model (MPDRM) that mitigates the negative impact of individual differences, enabling the model to focus on general and robust facial depression cues. The proposed model consists of three main components: the Multi-scale Depression Feature Network (MDFN), the Multi-scale Personality Feature Network (MPFN), and the Relational Attention Recognition Module (RARM). The MDFN extracts depression-related information, while the contrastive learning-based MPFN extracts stable personalized information. In both MDFN and MPFN, we insert the Multi-scale Motion Pattern Extraction Module (MMP) to capture rich multi-scale spatiotemporal facial features. Finally, the RARM is designed to enhance the representation of depression and output the results. Cross-validation on a specifically constructed longitudinal dataset demonstrates that our model outperforms other models. Experimental results indicate that suppressing personalized information of facial movements can effectively improve the accuracy of depression recognition.
AB - Automatic depression recognition based on facial movements in videos has become a research hotspot. However, existing methods tend to confuse individual inherent facial behavioral habits with characteristics specific to depression, which leads to misjudgments. To address this, we propose a Multi-scale Personalized Depression Recognition Model (MPDRM) that mitigates the negative impact of individual differences, enabling the model to focus on general and robust facial depression cues. The proposed model consists of three main components: the Multi-scale Depression Feature Network (MDFN), the Multi-scale Personality Feature Network (MPFN), and the Relational Attention Recognition Module (RARM). The MDFN extracts depression-related information, while the contrastive learning-based MPFN extracts stable personalized information. In both MDFN and MPFN, we insert the Multi-scale Motion Pattern Extraction Module (MMP) to capture rich multi-scale spatiotemporal facial features. Finally, the RARM is designed to enhance the representation of depression and output the results. Cross-validation on a specifically constructed longitudinal dataset demonstrates that our model outperforms other models. Experimental results indicate that suppressing personalized information of facial movements can effectively improve the accuracy of depression recognition.
KW - Automatic depression recognition
KW - Contrastive learning
KW - Facial movements
KW - Motion Pattern Extraction
KW - Personalized modeling
UR - http://www.scopus.com/inward/record.url?scp=85219037899&partnerID=8YFLogxK
U2 - 10.1016/j.neucom.2025.129669
DO - 10.1016/j.neucom.2025.129669
M3 - Article
AN - SCOPUS:85219037899
SN - 0925-2312
VL - 632
JO - Neurocomputing
JF - Neurocomputing
M1 - 129669
ER -