TY - JOUR
T1 - Distributed Stochastic Gradient Tracking Algorithm With Variance Reduction for Non-Convex Optimization
AU - Jiang, Xia
AU - Zeng, Xianlin
AU - Sun, Jian
AU - Chen, Jie
N1 - Publisher Copyright:
© 2012 IEEE.
PY - 2023/9/1
Y1 - 2023/9/1
N2 - This article proposes a distributed stochastic algorithm with variance reduction for general smooth non-convex finite-sum optimization, which has wide applications in signal processing and machine learning communities. In distributed setting, a large number of samples are allocated to multiple agents in the network. Each agent computes local stochastic gradient and communicates with its neighbors to seek for the global optimum. In this article, we develop a modified variance reduction technique to deal with the variance introduced by stochastic gradients. Combining gradient tracking and variance reduction techniques, this article proposes a distributed stochastic algorithm, gradient tracking algorithm with variance reduction (GT-VR), to solve large-scale non-convex finite-sum optimization over multiagent networks. A complete and rigorous proof shows that the GT-VR algorithm converges to the first-order stationary points with O(1/k) convergence rate. In addition, we provide the complexity analysis of the proposed algorithm. Compared with some existing first-order methods, the proposed algorithm has a lower O(PMϵ-1) gradient complexity under some mild condition. By comparing state-of-the-art algorithms and GT-VR in numerical simulations, we verify the efficiency of the proposed algorithm.
AB - This article proposes a distributed stochastic algorithm with variance reduction for general smooth non-convex finite-sum optimization, which has wide applications in signal processing and machine learning communities. In distributed setting, a large number of samples are allocated to multiple agents in the network. Each agent computes local stochastic gradient and communicates with its neighbors to seek for the global optimum. In this article, we develop a modified variance reduction technique to deal with the variance introduced by stochastic gradients. Combining gradient tracking and variance reduction techniques, this article proposes a distributed stochastic algorithm, gradient tracking algorithm with variance reduction (GT-VR), to solve large-scale non-convex finite-sum optimization over multiagent networks. A complete and rigorous proof shows that the GT-VR algorithm converges to the first-order stationary points with O(1/k) convergence rate. In addition, we provide the complexity analysis of the proposed algorithm. Compared with some existing first-order methods, the proposed algorithm has a lower O(PMϵ-1) gradient complexity under some mild condition. By comparing state-of-the-art algorithms and GT-VR in numerical simulations, we verify the efficiency of the proposed algorithm.
KW - Complexity analysis
KW - distributed algorithm
KW - non-convex finite-sum optimization
KW - stochastic gradient
KW - variance reduction
UR - http://www.scopus.com/inward/record.url?scp=85132509520&partnerID=8YFLogxK
U2 - 10.1109/TNNLS.2022.3170944
DO - 10.1109/TNNLS.2022.3170944
M3 - Article
C2 - 35536804
AN - SCOPUS:85132509520
SN - 2162-237X
VL - 34
SP - 5310
EP - 5321
JO - IEEE Transactions on Neural Networks and Learning Systems
JF - IEEE Transactions on Neural Networks and Learning Systems
IS - 9
ER -