TY - JOUR
T1 - On Consensus-Optimality Trade-offs in Collaborative Deep Learning
AU - Jiang, Zhanhong
AU - Balu, Aditya
AU - Hegde, Chinmay
AU - Sarkar, Soumik
N1 - Publisher Copyright:
© Copyright © 2021 Jiang, Balu, Hegde and Sarkar.
PY - 2021/9/14
Y1 - 2021/9/14
N2 - In distributed machine learning, where agents collaboratively learn from diverse private data sets, there is a fundamental tension between consensus and optimality. In this paper, we build on recent algorithmic progresses in distributed deep learning to explore various consensus-optimality trade-offs over a fixed communication topology. First, we propose the incremental consensus-based distributed stochastic gradient descent (i-CDSGD) algorithm, which involves multiple consensus steps (where each agent communicates information with its neighbors) within each SGD iteration. Second, we propose the generalized consensus-based distributed SGD (g-CDSGD) algorithm that enables us to navigate the full spectrum from complete consensus (all agents agree) to complete disagreement (each agent converges to individual model parameters). We analytically establish convergence of the proposed algorithms for strongly convex and nonconvex objective functions; we also analyze the momentum variants of the algorithms for the strongly convex case. We support our algorithms via numerical experiments, and demonstrate significant improvements over existing methods for collaborative deep learning.
AB - In distributed machine learning, where agents collaboratively learn from diverse private data sets, there is a fundamental tension between consensus and optimality. In this paper, we build on recent algorithmic progresses in distributed deep learning to explore various consensus-optimality trade-offs over a fixed communication topology. First, we propose the incremental consensus-based distributed stochastic gradient descent (i-CDSGD) algorithm, which involves multiple consensus steps (where each agent communicates information with its neighbors) within each SGD iteration. Second, we propose the generalized consensus-based distributed SGD (g-CDSGD) algorithm that enables us to navigate the full spectrum from complete consensus (all agents agree) to complete disagreement (each agent converges to individual model parameters). We analytically establish convergence of the proposed algorithms for strongly convex and nonconvex objective functions; we also analyze the momentum variants of the algorithms for the strongly convex case. We support our algorithms via numerical experiments, and demonstrate significant improvements over existing methods for collaborative deep learning.
KW - collaborative deep learning
KW - consensus-optimality
KW - convergence
KW - distributed optimization
KW - sgd
UR - http://www.scopus.com/inward/record.url?scp=85117956596&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85117956596&partnerID=8YFLogxK
U2 - 10.3389/frai.2021.573731
DO - 10.3389/frai.2021.573731
M3 - Article
AN - SCOPUS:85117956596
SN - 2624-8212
VL - 4
JO - Frontiers in Artificial Intelligence
JF - Frontiers in Artificial Intelligence
M1 - 573731
ER -