Dynamic threshold integrate and fire neuron model for low latency spiking neural networks

Xiyan Wu, Yufei Zhao, Yong Song*, Yurong Jiang, Yashuo Bai, Xinyi Li, Ya Zhou, Xin Yang, Qun Hao

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

5 引用 (Scopus)

摘要

Spiking Neural Networks (SNNs) operate with asynchronous discrete events which enable lower power and greater computational efficiency on event-driven hardware than Artificial Neural Networks (ANNs). Conventional ANN-to-SNN conversion methods usually employ Integrate and Fire (IF) neuron model with a fixed threshold to act as Rectified Linear Unit (ReLU). However, there is a large demand for the input spikes to reach the fixed threshold and fire, which leads to high inference latency. In this work, we propose a Dynamic Threshold Integrate and Fire (DTIF) neuron model by exploiting the biological neuron threshold variability, where the threshold is inversely related to the neuron input. The spike activity is increased by dynamically adjusting the threshold at each simulation time-step to reduce the latency. Compared to the state-of-the-art conversion methods, the ANN-to-SNN conversion using DTIF model has lower latency with competitive accuracy, which has been verified by deep architecture on image classification tasks including MNIST, CAIFAR-10, and CIFAR-100 datasets. Moreover, it achieves 7.14 × faster inference under 0.44 × energy consumption than the typical method of maximum normalization.

源语言英语
文章编号126247
期刊Neurocomputing
544
DOI
出版状态已出版 - 1 8月 2023

指纹

探究 'Dynamic threshold integrate and fire neuron model for low latency spiking neural networks' 的科研主题。它们共同构成独一无二的指纹。

引用此