Dataflow optimization with layer-wise design variables estimation method for enflame CNN accelerators

Tian Chen, Yu an Tan, Zheng Zhang, Nan Luo, Bin Li, Yuanzhang Li*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

As convolution layers have been proved to be the most time-consuming operation in convolutional neural network (CNN) algorithms, many efficient CNN accelerators have been designed to boost the performance of convolution operations. Previous works on CNN acceleration usually use fixed design variables for diverse convolutional layers, which would lead to inefficient data movements and low utilization of computing resource. We tackle this issue by proposing a flexible dataflow optimization method with design variables estimation for different layers. The optimization method first narrows the design space by the priori constraints, and then enumerates all legal solutions to select the optimal design variables. We demonstrate the effectiveness of the proposed optimization method by implementing representative CNN models (VGG-16, ResNet-18 and MobileNet V1) on Enflame Technology's programmable CNN accelerator, General Computing Unit (GCU). The results indicate that our optimization can significantly enhance the throughput of the convolution layers in ResNet, VGG and MobileNet on GCU, with improvement of up to 1.84×. Furthermore, it achieves up to 2.08× of GCU utilization specifically for the convolution layers of ResNet on GCU.

Original languageEnglish
Article number104869
JournalJournal of Parallel and Distributed Computing
Volume189
DOIs
Publication statusPublished - Jul 2024

Keywords

  • Convolutional neural networks (CNNs)
  • General computing unit (GCU)
  • Optimization
  • Programmable dataflow

Fingerprint

Dive into the research topics of 'Dataflow optimization with layer-wise design variables estimation method for enflame CNN accelerators'. Together they form a unique fingerprint.

Cite this