TY - JOUR
T1 - Random sketch learning for deep neural networks in edge computing
AU - Li, Bin
AU - Chen, Peijun
AU - Liu, Hongfu
AU - Guo, Weisi
AU - Cao, Xianbin
AU - Du, Junzhao
AU - Zhao, Chenglin
AU - Zhang, Jun
N1 - Publisher Copyright:
© 2021, The Author(s), under exclusive licence to Springer Nature America, Inc.
PY - 2021/3
Y1 - 2021/3
N2 - Despite the great potential of deep neural networks (DNNs), they require massive weights and huge computational resources, creating a vast gap when deploying artificial intelligence at low-cost edge devices. Current lightweight DNNs, achieved by high-dimensional space pre-training and post-compression, present challenges when covering the resources deficit, making tiny artificial intelligence hard to be implemented. Here we report an architecture named random sketch learning, or Rosler, for computationally efficient tiny artificial intelligence. We build a universal compressing-while-training framework that directly learns a compact model and, most importantly, enables computationally efficient on-device learning. As validated on different models and datasets, it attains substantial memory reduction of ~50–90× (16-bits quantization), compared with fully connected DNNs. We demonstrate it on low-cost hardware, whereby the computation is accelerated by >180× and the energy consumption is reduced by ~10×. Our method paves the way for deploying tiny artificial intelligence in many scientific and industrial applications.
AB - Despite the great potential of deep neural networks (DNNs), they require massive weights and huge computational resources, creating a vast gap when deploying artificial intelligence at low-cost edge devices. Current lightweight DNNs, achieved by high-dimensional space pre-training and post-compression, present challenges when covering the resources deficit, making tiny artificial intelligence hard to be implemented. Here we report an architecture named random sketch learning, or Rosler, for computationally efficient tiny artificial intelligence. We build a universal compressing-while-training framework that directly learns a compact model and, most importantly, enables computationally efficient on-device learning. As validated on different models and datasets, it attains substantial memory reduction of ~50–90× (16-bits quantization), compared with fully connected DNNs. We demonstrate it on low-cost hardware, whereby the computation is accelerated by >180× and the energy consumption is reduced by ~10×. Our method paves the way for deploying tiny artificial intelligence in many scientific and industrial applications.
UR - http://www.scopus.com/inward/record.url?scp=85107531510&partnerID=8YFLogxK
U2 - 10.1038/s43588-021-00039-6
DO - 10.1038/s43588-021-00039-6
M3 - Article
AN - SCOPUS:85107531510
SN - 2662-8457
VL - 1
SP - 221
EP - 228
JO - Nature Computational Science
JF - Nature Computational Science
IS - 3
ER -