Accelerating Deep Learning Systems via Critical Set Identification and Model Compression

Rui Han, Chi Harold Liu*, Shilin Li, Shilin Wen, Xue Liu

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

12 引用 (Scopus)

摘要

Modern distributed engines are increasingly deployed to accelerate large-scaled deep learning (DL) training jobs. While the parallelism of distributed workers/nodes promises the scalability, the computation and communication overheads of the underlying iterative solving algorithms, e.g., stochastic gradient decent, unfortunately become the bottleneck for distributed DL training jobs. Existing approaches address such limitations by designing more efficient synchronization algorithms and model compressing techniques, but do not adequately address issues relating to processing massive datasets. In this article, we propose ClipDL, which accelerates the deep learning systems by simultaneously decreasing the number of model parameters as well as reducing the computations on critical data only. The core component of ClipDL is the estimation of critical set based on the observation that large proportions of input data have little influence on model parameter updating in many prevalent DL algorithms. We implemented ClipDL on Spark (a popular distributed engine for big data) and BigDL (based on de-factor distributed DL training architecture, parameter server), and integrated it with representative model compression techniques. The exhaustive experiments on real DL applications and datasets show ClipDL accelerates model training process by an average of 2.32 times while only incurring accuracy losses of 1.86 percent.

源语言英语
文章编号8977355
页(从-至)1059-1070
页数12
期刊IEEE Transactions on Computers
69
7
DOI
出版状态已出版 - 1 7月 2020

指纹

探究 'Accelerating Deep Learning Systems via Critical Set Identification and Model Compression' 的科研主题。它们共同构成独一无二的指纹。

引用此