Abstract
Electroencephalogram(EEG)-based emotion recognition is crucial for advancing human-computer interaction (HCI), and brain network features have become a key research focus. While existing methods often concatenate brain network features with traditional single-channel features to enhance recognition performance, this direct concatenation undermines the spatial information of brain networks and hinders effective application of deep learning. In this work, we propose a novel feature fusion strategy that effectively combines two-dimensional brain effective connectivity (BEC) network features with one-dimensional spectral power features while preserving spatial information. To leverage the spatial topological properties of brain networks and the one-dimensional correlations in fused features, we further introduce a Dual-channel 1D-CNN based on Spatially Unidimensional Self-Attention (SAD-1D-CNN), designed to extract discriminative features by capturing spatial correlations within the combined data. Results show 90.61% accuracy on SEED and 82.13% on SEED-IV (2.68% higher than state-of-the-art). Comprehensive tests confirm the superiority of our fusion strategy and SAD-1D-CNN in emotion recognition. Parameter visualization reveals the attention module’s ability to automatically focus on emotion-related core brain regions, and ablation experiments validate the necessity of each network module. These findings offer new perspectives for advancing emotion recognition research.
| Original language | English |
|---|---|
| Article number | 175 |
| Journal | Cognitive Neurodynamics |
| Volume | 19 |
| Issue number | 1 |
| DOIs | |
| Publication status | Published - Dec 2025 |
| Externally published | Yes |
Keywords
- 1D-CNN
- Brain effective connectivity (BEC)
- Emotion recognition
- Feature fusion
- Self-Attention