TY - JOUR
T1 - BrainyHGNN
T2 - Brain-Inspired Memory Retrieval and Cross-modal Interaction for Emotion Recognition in Conversations
AU - Wang, Qixin
AU - Li, Ziyu
AU - Li, Xiuxing
AU - Jia, Tianyuan
AU - Li, Qing
AU - Yao, Li
AU - Wu, Xia
N1 - Publisher Copyright:
© 1991-2012 IEEE.
PY - 2025
Y1 - 2025
N2 - Research on emotion recognition in conversations emphasises the importance of complex relationships between conversational context and multimodality. Graph-based methods, particularly hypergraph-based methods have shown promise in capturing these relationships. However, challenges persist in avoiding redundant context while capturing essential information for optimal context embeddings and fully leveraging cross-modal complementarities for sufficient fusion. In contrast, the human brain flexibly retrieves relevant memories and integrates multi-modal data for accurate recognition. Based on this superiority, we propose BrainyHGNN, a brain-inspired hypergraph neural network. It integrates a Dynamic Memory Selector for contextual hyperedges, mimicking selective memory retrieval mechanisms for adaptive and modality-specific context retrieval. HierSensNet is designed for multi-modal hyperedges, mirroring hierarchical cross-modal interaction mechanisms to ensure effective multimodal fusion. Experimental results on two benchmark datasets validate the superior performance of BrainyHGNN, confirming the effectiveness of its innovative approach. This work highlights the potential of brain-inspired methods to advance flexible context retrieval and sufficient multimodal fusion, presenting a promising direction for future research in this domain.
AB - Research on emotion recognition in conversations emphasises the importance of complex relationships between conversational context and multimodality. Graph-based methods, particularly hypergraph-based methods have shown promise in capturing these relationships. However, challenges persist in avoiding redundant context while capturing essential information for optimal context embeddings and fully leveraging cross-modal complementarities for sufficient fusion. In contrast, the human brain flexibly retrieves relevant memories and integrates multi-modal data for accurate recognition. Based on this superiority, we propose BrainyHGNN, a brain-inspired hypergraph neural network. It integrates a Dynamic Memory Selector for contextual hyperedges, mimicking selective memory retrieval mechanisms for adaptive and modality-specific context retrieval. HierSensNet is designed for multi-modal hyperedges, mirroring hierarchical cross-modal interaction mechanisms to ensure effective multimodal fusion. Experimental results on two benchmark datasets validate the superior performance of BrainyHGNN, confirming the effectiveness of its innovative approach. This work highlights the potential of brain-inspired methods to advance flexible context retrieval and sufficient multimodal fusion, presenting a promising direction for future research in this domain.
KW - Brain-inspired methods
KW - emotion recognition in conversations
KW - hypergraph neural networks
KW - multimodal fusion
UR - http://www.scopus.com/inward/record.url?scp=105005870159&partnerID=8YFLogxK
U2 - 10.1109/TCSVT.2025.3572226
DO - 10.1109/TCSVT.2025.3572226
M3 - Article
AN - SCOPUS:105005870159
SN - 1051-8215
JO - IEEE Transactions on Circuits and Systems for Video Technology
JF - IEEE Transactions on Circuits and Systems for Video Technology
ER -