1000× Faster Camera and Machine Vision with Ordinary Devices

Tiejun Huang, Yajing Zheng, Zhaofei Yu*, Rui Chen, Yuan Li, Ruiqin Xiong, Lei Ma, Junwei Zhao, Siwei Dong, Lin Zhu, Jianing Li, Shanshan Jia, Yihua Fu, Boxin Shi, Si Wu, Yonghong Tian

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

46 引用 (Scopus)
Plum Print visual indicator of research metrics
  • Citations
    • Citation Indexes: 41
  • Captures
    • Readers: 27
  • Mentions
    • News Mentions: 4
see details

摘要

In digital cameras, we find a major limitation: the image and video form inherited from a film camera obstructs it from capturing the rapidly changing photonic world. Here, we present vform, a bit sequence array where each bit represents whether the accumulation of photons has reached a threshold, to record and reconstruct the scene radiance at any moment. By employing only consumer-level complementary metal–oxide semiconductor (CMOS) sensors and integrated circuits, we have developed a spike camera that is 1000× faster than conventional cameras. By treating vform as spike trains in biological vision, we have further developed a spiking neural network (SNN)-based machine vision system that combines the speed of the machine and the mechanism of biological vision, achieving high-speed object detection and tracking 1000× faster than human vision. We demonstrate the utility of the spike camera and the super vision system in an assistant referee and target pointing system. Our study is expected to fundamentally revolutionize the image and video concepts and related industries, including photography, movies, and visual media, and to unseal a new SNN-enabled speed-free machine vision era.

源语言英语
页(从-至)110-119
页数10
期刊Engineering
25
DOI
出版状态已出版 - 6月 2023
已对外发布

指纹

探究 '1000× Faster Camera and Machine Vision with Ordinary Devices' 的科研主题。它们共同构成独一无二的指纹。

引用此

Huang, T., Zheng, Y., Yu, Z., Chen, R., Li, Y., Xiong, R., Ma, L., Zhao, J., Dong, S., Zhu, L., Li, J., Jia, S., Fu, Y., Shi, B., Wu, S., & Tian, Y. (2023). 1000× Faster Camera and Machine Vision with Ordinary Devices. Engineering, 25, 110-119. https://doi.org/10.1016/j.eng.2022.01.012