TY - GEN
T1 - Binary Separable Convolutional
T2 - 2019 IEEE International Conference on Signal, Information and Data Processing, ICSIDP 2019
AU - Jing, Donglin
AU - Tang, Linbo
AU - Pan, Yu
AU - Tang, Wei
AU - Zhou, Shichao
N1 - Publisher Copyright:
© 2019 IEEE.
PY - 2019/12
Y1 - 2019/12
N2 - The training and running of neural network require large computational space and memory space, which makes it difficult to deploy on a resource-constrained embedded sys-tems. To address this limitation, we introduce a two-stage pipeline: Depth separable, local binary. Our method is divided into two steps. Firstly, a deep separable convolution is applied to each input channel. Then, point-by-point convolution is applied to the feature map obtained by the filter. In the second step, we use local binarization method to initialize the filter corresponding to the input channel into sparse binary. In network training, the sparse binary filter remains fixed and only needs to train convolution of 1×1 size. Our experimental results show that, on the basis of approximate accuracy with the original network, we have reduced the number of convolution parameters by 9x to 10x, and reduced the training time and testing time to by 2x. Our compression method helps to deploy complex neural networks on resource-constrained embedded platform.
AB - The training and running of neural network require large computational space and memory space, which makes it difficult to deploy on a resource-constrained embedded sys-tems. To address this limitation, we introduce a two-stage pipeline: Depth separable, local binary. Our method is divided into two steps. Firstly, a deep separable convolution is applied to each input channel. Then, point-by-point convolution is applied to the feature map obtained by the filter. In the second step, we use local binarization method to initialize the filter corresponding to the input channel into sparse binary. In network training, the sparse binary filter remains fixed and only needs to train convolution of 1×1 size. Our experimental results show that, on the basis of approximate accuracy with the original network, we have reduced the number of convolution parameters by 9x to 10x, and reduced the training time and testing time to by 2x. Our compression method helps to deploy complex neural networks on resource-constrained embedded platform.
KW - Compression
KW - Convolutionl Neural Network
KW - Image Classification
KW - Local Binary
KW - Separable
UR - http://www.scopus.com/inward/record.url?scp=85091936810&partnerID=8YFLogxK
U2 - 10.1109/ICSIDP47821.2019.9173414
DO - 10.1109/ICSIDP47821.2019.9173414
M3 - Conference contribution
AN - SCOPUS:85091936810
T3 - ICSIDP 2019 - IEEE International Conference on Signal, Information and Data Processing 2019
BT - ICSIDP 2019 - IEEE International Conference on Signal, Information and Data Processing 2019
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 11 December 2019 through 13 December 2019
ER -