Sign Language Gesture Recognition and Classification Based on Event Camera with Spiking Neural Networks

Xuena Chen, Li Su*, Jinxiu Zhao, Keni Qiu, Na Jiang, Guang Zhai

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

18 引用 (Scopus)

摘要

Sign language recognition has been utilized in human–machine interactions, improving the lives of people with speech impairments or who rely on nonverbal instructions. Thanks to its higher temporal resolution, less visual redundancy information and lower energy consumption, the use of an event camera with a new dynamic vision sensor (DVS) shows promise with regard to sign language recognition with robot perception and intelligent control. Although previous work has focused on event camera-based, simple gesture datasets, such as DVS128Gesture, event camera gesture datasets inspired by sign language are critical, which poses a great impediment to the development of event camera-based sign language recognition. An effective method to extract spatio-temporal features from event data is significantly desired. Firstly, the event-based sign language gesture datasets are proposed and the data have two sources: traditional sign language videos to event stream (DVS_Sign_v2e) and DAVIS346 (DVS_Sign). In the present dataset, data are divided into five classification, verbs, quantifiers, position, things and people, adapting to actual scenarios where robots provide instruction or assistance. Sign language classification is demonstrated in spike neuron networks with a spatio-temporal back-propagation training method, leading to the best recognition accuracy of 77%. This work paves the way for the combination of event camera-based sign language gesture recognition and robotic perception for the future intelligent systems.

源语言英语
文章编号786
期刊Electronics (Switzerland)
12
4
DOI
出版状态已出版 - 2月 2023

指纹

探究 'Sign Language Gesture Recognition and Classification Based on Event Camera with Spiking Neural Networks' 的科研主题。它们共同构成独一无二的指纹。

引用此