Efficient and Structural Gradient Compression with Principal Component Analysis for Distributed Training

Jiaxin Tan, Chao Yao, Zehua Guo*

*此作品的通讯作者

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

Distributed machine learning is a promising machine learning approach for academia and industry. It can generate a machine learning model for dispersed training data via iterative training in a distributed fashion. To speed up the training process of distributed machine learning, it is essential to reduce the communication load among training nodes. In this paper, we propose a layer-wise gradient compression scheme based on principal component analysis and error accumulation. The key of our solution is to consider the gradient characteristics and architecture of neural networks by taking advantage of the compression ability enabled by PCA and the feedback ability enabled by error accumulation. The preliminary results on image classification task show that our scheme achieves good performance and reduces 97% of the gradient transmission.

源语言英语
主期刊名Proceedings of the 7th Asia-Pacific Workshop on Networking, APNET 2023
出版商Association for Computing Machinery, Inc
217-218
页数2
ISBN(电子版)9798400707827
DOI
出版状态已出版 - 29 6月 2023
活动7th Asia-Pacific Workshop on Networking, APNET 2023 - Hong Kong, 中国
期限: 29 6月 202330 6月 2023

出版系列

姓名Proceedings of the 7th Asia-Pacific Workshop on Networking, APNET 2023

会议

会议7th Asia-Pacific Workshop on Networking, APNET 2023
国家/地区中国
Hong Kong
时期29/06/2330/06/23

指纹

探究 'Efficient and Structural Gradient Compression with Principal Component Analysis for Distributed Training' 的科研主题。它们共同构成独一无二的指纹。

引用此