TY - GEN
T1 - Competitive and Collaborative Learning Accelerates the Convergence of Deep Convolutional Neural Networks
AU - Dang, Yanbin
AU - Yang, Yuliang
AU - Chen, Yueyun
AU - Zhu, Mengyu
AU - Yin, Dehui
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - In the training of convolutional neural networks (CNNs), the layer-by-layer learning based on the backpropagation (BP) algorithm causes that in each round of weights update, the learning of the latter layer determines the learning of the former layer, while the former layer cannot directly affect the latter layer. This means that the flow of error information is unidirectional, causing non-cooperative learning between layers, thereby reducing the convergence speed of the networks. In this work, we propose a network structure that evaluates the relative contribution of each layer in the CNNs to the final output error. During training, it indirectly realizes the bidirectional flow of information between layers, achieving the purpose of cross-layer collaborative learning. Our algorithm also fuses features at different scales on the detection networks, which we call the flexible feature fusion network(FFN). On public datasets, we have conducted rich experiments. With the help of FFN, the convergence speed of the object detection model is greatly improved. Without pre-training weight initialization, the convergence speed of the model is approximately doubled.
AB - In the training of convolutional neural networks (CNNs), the layer-by-layer learning based on the backpropagation (BP) algorithm causes that in each round of weights update, the learning of the latter layer determines the learning of the former layer, while the former layer cannot directly affect the latter layer. This means that the flow of error information is unidirectional, causing non-cooperative learning between layers, thereby reducing the convergence speed of the networks. In this work, we propose a network structure that evaluates the relative contribution of each layer in the CNNs to the final output error. During training, it indirectly realizes the bidirectional flow of information between layers, achieving the purpose of cross-layer collaborative learning. Our algorithm also fuses features at different scales on the detection networks, which we call the flexible feature fusion network(FFN). On public datasets, we have conducted rich experiments. With the help of FFN, the convergence speed of the object detection model is greatly improved. Without pre-training weight initialization, the convergence speed of the model is approximately doubled.
KW - collaborative learning
KW - multi-scale feature fusion
KW - object detection
KW - training efficiency
UR - http://www.scopus.com/inward/record.url?scp=85132208343&partnerID=8YFLogxK
U2 - 10.1109/ICCCBDA55098.2022.9778930
DO - 10.1109/ICCCBDA55098.2022.9778930
M3 - Conference contribution
AN - SCOPUS:85132208343
T3 - 2022 7th International Conference on Cloud Computing and Big Data Analytics, ICCCBDA 2022
SP - 431
EP - 438
BT - 2022 7th International Conference on Cloud Computing and Big Data Analytics, ICCCBDA 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 7th International Conference on Cloud Computing and Big Data Analytics, ICCCBDA 2022
Y2 - 22 April 2022 through 24 April 2022
ER -