Fast Distributed Gradient Methods

We study distributed optimization problems when $N$ nodes minimize the sum of their individual costs subject to a common vector variable. The costs are convex, have Lipschitz continuous gradient (with constant $L$), and bounded gradient. We propose two fast distributed gradient algorithms based on t...

Full description

Saved in:
Bibliographic Details
Main Authors Jakovetic, Dusan, Xavier, Joao, Moura, Jose M. F
Format Journal Article
LanguageEnglish
Published 13.12.2011
Subjects
Online AccessGet full text
DOI10.48550/arxiv.1112.2972

Cover

Abstract We study distributed optimization problems when $N$ nodes minimize the sum of their individual costs subject to a common vector variable. The costs are convex, have Lipschitz continuous gradient (with constant $L$), and bounded gradient. We propose two fast distributed gradient algorithms based on the centralized Nesterov gradient algorithm and establish their convergence rates in terms of the per-node communications $\mathcal{K}$ and the per-node gradient evaluations $k$. Our first method, Distributed Nesterov Gradient, achieves rates $O\left({\log \mathcal{K}}/{\mathcal{K}}\right)$ and $O\left({\log k}/{k}\right)$. Our second method, Distributed Nesterov gradient with Consensus iterations, assumes at all nodes knowledge of $L$ and $\mu(W)$ -- the second largest singular value of the $N \times N$ doubly stochastic weight matrix $W$. It achieves rates $O\left({1}/{\mathcal{K}^{2-\xi}}\right)$ and $O\left({1}/{k^2}\right)$ ($\xi>0$ arbitrarily small). Further, we give with both methods explicit dependence of the convergence constants on $N$ and $W$. Simulation examples illustrate our findings.
AbstractList We study distributed optimization problems when $N$ nodes minimize the sum of their individual costs subject to a common vector variable. The costs are convex, have Lipschitz continuous gradient (with constant $L$), and bounded gradient. We propose two fast distributed gradient algorithms based on the centralized Nesterov gradient algorithm and establish their convergence rates in terms of the per-node communications $\mathcal{K}$ and the per-node gradient evaluations $k$. Our first method, Distributed Nesterov Gradient, achieves rates $O\left({\log \mathcal{K}}/{\mathcal{K}}\right)$ and $O\left({\log k}/{k}\right)$. Our second method, Distributed Nesterov gradient with Consensus iterations, assumes at all nodes knowledge of $L$ and $\mu(W)$ -- the second largest singular value of the $N \times N$ doubly stochastic weight matrix $W$. It achieves rates $O\left({1}/{\mathcal{K}^{2-\xi}}\right)$ and $O\left({1}/{k^2}\right)$ ($\xi>0$ arbitrarily small). Further, we give with both methods explicit dependence of the convergence constants on $N$ and $W$. Simulation examples illustrate our findings.
Author Moura, Jose M. F
Jakovetic, Dusan
Xavier, Joao
Author_xml – sequence: 1
  givenname: Dusan
  surname: Jakovetic
  fullname: Jakovetic, Dusan
– sequence: 2
  givenname: Joao
  surname: Xavier
  fullname: Xavier, Joao
– sequence: 3
  givenname: Jose M. F
  surname: Moura
  fullname: Moura, Jose M. F
BackLink https://doi.org/10.48550/arXiv.1112.2972$$DView paper in arXiv
BookMark eNrjYmDJy89LZWCQMDTQM7EwNTXQTyyqyCzTMzQ0NNIzsjQ34mRQdEssLlFwySwuKcpMKi1JTVFwL0pMyUzNK1HwTS3JyE8p5mFgTUvMKU7lhdLcDHJuriHOHrpgs-ILijJzE4sq40FmxoPMNCaoAACzzyw-
ContentType Journal Article
Copyright http://arxiv.org/licenses/nonexclusive-distrib/1.0
Copyright_xml – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0
DBID AKY
AKZ
GOX
DOI 10.48550/arxiv.1112.2972
DatabaseName arXiv Computer Science
arXiv Mathematics
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 1112_2972
GroupedDBID AKY
AKZ
GOX
ID FETCH-arxiv_primary_1112_29723
IEDL.DBID GOX
IngestDate Tue Jul 22 23:03:25 EDT 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-arxiv_primary_1112_29723
OpenAccessLink https://arxiv.org/abs/1112.2972
ParticipantIDs arxiv_primary_1112_2972
PublicationCentury 2000
PublicationDate 2011-12-13
PublicationDateYYYYMMDD 2011-12-13
PublicationDate_xml – month: 12
  year: 2011
  text: 2011-12-13
  day: 13
PublicationDecade 2010
PublicationYear 2011
Score 2.9473293
SecondaryResourceType preprint
Snippet We study distributed optimization problems when $N$ nodes minimize the sum of their individual costs subject to a common vector variable. The costs are convex,...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Information Theory
Mathematics - Information Theory
Title Fast Distributed Gradient Methods
URI https://arxiv.org/abs/1112.2972
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwY2BQMTQ3TDM1NTbVBcZoiq4J6LChpESjRN2U1CRLE1NTYGkIvhvQ18_MI9TEK8I0golBHrYXJrGoIrMMcj5wUrE-MCMa6RlZmgPLWGYjI1Dfyt0_AjLZCD6JC6ocrgzYwgSLIFURboIM_NC2nYIjJDKEGJhS80QYFN0Si0sUXEDn04KulkpNUXAvAi-zKlHwBd_eXCzKIOfmGuLsoQs2NL4Acv4DqNFsFA-yzliMgQXYSU-VYFBITU4zS7FIMTVPs7QwSTUxS0o2NQQWg6kGqSnAFoaZmSSDOA5DpHDKSDNwgUcwDY10DY1lGFhKikpTZYFVYEmSHDggAHWCX1M
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Fast+Distributed+Gradient+Methods&rft.au=Jakovetic%2C+Dusan&rft.au=Xavier%2C+Joao&rft.au=Moura%2C+Jose+M.+F&rft.date=2011-12-13&rft_id=info:doi/10.48550%2Farxiv.1112.2972&rft.externalDocID=1112_2972