TY - JOUR
T1 - STCA-SNN
T2 - self-attention-based temporal-channel joint attention for spiking neural networks
AU - Wu, Xiyan
AU - Song, Yong
AU - Zhou, Ya
AU - Jiang, Yurong
AU - Bai, Yashuo
AU - Li, Xinyi
AU - Yang, Xin
N1 - Publisher Copyright:
Copyright © 2023 Wu, Song, Zhou, Jiang, Bai, Li and Yang.
PY - 2023
Y1 - 2023
N2 - Spiking Neural Networks (SNNs) have shown great promise in processing spatio-temporal information compared to Artificial Neural Networks (ANNs). However, there remains a performance gap between SNNs and ANNs, which impedes the practical application of SNNs. With intrinsic event-triggered property and temporal dynamics, SNNs have the potential to effectively extract spatio-temporal features from event streams. To leverage the temporal potential of SNNs, we propose a self-attention-based temporal-channel joint attention SNN (STCA-SNN) with end-to-end training, which infers attention weights along both temporal and channel dimensions concurrently. It models global temporal and channel information correlations with self-attention, enabling the network to learn ‘what’ and ‘when’ to attend simultaneously. Our experimental results show that STCA-SNNs achieve better performance on N-MNIST (99.67%), CIFAR10-DVS (81.6%), and N-Caltech 101 (80.88%) compared with the state-of-the-art SNNs. Meanwhile, our ablation study demonstrates that STCA-SNNs improve the accuracy of event stream classification tasks.
AB - Spiking Neural Networks (SNNs) have shown great promise in processing spatio-temporal information compared to Artificial Neural Networks (ANNs). However, there remains a performance gap between SNNs and ANNs, which impedes the practical application of SNNs. With intrinsic event-triggered property and temporal dynamics, SNNs have the potential to effectively extract spatio-temporal features from event streams. To leverage the temporal potential of SNNs, we propose a self-attention-based temporal-channel joint attention SNN (STCA-SNN) with end-to-end training, which infers attention weights along both temporal and channel dimensions concurrently. It models global temporal and channel information correlations with self-attention, enabling the network to learn ‘what’ and ‘when’ to attend simultaneously. Our experimental results show that STCA-SNNs achieve better performance on N-MNIST (99.67%), CIFAR10-DVS (81.6%), and N-Caltech 101 (80.88%) compared with the state-of-the-art SNNs. Meanwhile, our ablation study demonstrates that STCA-SNNs improve the accuracy of event stream classification tasks.
KW - event streams
KW - neuromorphic computing
KW - self-attention
KW - spiking neural networks
KW - temporal-channel
UR - http://www.scopus.com/inward/record.url?scp=85177582914&partnerID=8YFLogxK
U2 - 10.3389/fnins.2023.1261543
DO - 10.3389/fnins.2023.1261543
M3 - Article
AN - SCOPUS:85177582914
SN - 1662-4548
VL - 17
JO - Frontiers in Neuroscience
JF - Frontiers in Neuroscience
M1 - 1261543
ER -