Efficient and Structural Gradient Compression with Principal Component Analysis for Distributed Training

Jiaxin Tan, Chao Yao, Zehua Guo*

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Distributed machine learning is a promising machine learning approach for academia and industry. It can generate a machine learning model for dispersed training data via iterative training in a distributed fashion. To speed up the training process of distributed machine learning, it is essential to reduce the communication load among training nodes. In this paper, we propose a layer-wise gradient compression scheme based on principal component analysis and error accumulation. The key of our solution is to consider the gradient characteristics and architecture of neural networks by taking advantage of the compression ability enabled by PCA and the feedback ability enabled by error accumulation. The preliminary results on image classification task show that our scheme achieves good performance and reduces 97% of the gradient transmission.

Original languageEnglish
Title of host publicationProceedings of the 7th Asia-Pacific Workshop on Networking, APNET 2023
PublisherAssociation for Computing Machinery, Inc
Pages217-218
Number of pages2
ISBN (Electronic)9798400707827
DOIs
Publication statusPublished - 29 Jun 2023
Event7th Asia-Pacific Workshop on Networking, APNET 2023 - Hong Kong, China
Duration: 29 Jun 202330 Jun 2023

Publication series

NameProceedings of the 7th Asia-Pacific Workshop on Networking, APNET 2023

Conference

Conference7th Asia-Pacific Workshop on Networking, APNET 2023
Country/TerritoryChina
CityHong Kong
Period29/06/2330/06/23

Fingerprint

Dive into the research topics of 'Efficient and Structural Gradient Compression with Principal Component Analysis for Distributed Training'. Together they form a unique fingerprint.

Cite this