Warning: Full texts from electronic resources are only available from the university network. You are currently outside this network. Please log in to access full texts.
Fast Distributed Gradient Methods
We study distributed optimization problems when $N$ nodes minimize the sum of their individual costs subject to a common vector variable. The costs are convex, have Lipschitz continuous gradient (with constant $L$), and bounded gradient. We propose two fast distributed gradient algorithms based on t...
Saved in:
Main Authors | , , |
---|---|
Format | Journal Article |
Language | English |
Published |
13.12.2011
|
Subjects | |
Online Access | Get full text |
DOI | 10.48550/arxiv.1112.2972 |
Cover
Abstract | We study distributed optimization problems when $N$ nodes minimize the sum of
their individual costs subject to a common vector variable. The costs are
convex, have Lipschitz continuous gradient (with constant $L$), and bounded
gradient. We propose two fast distributed gradient algorithms based on the
centralized Nesterov gradient algorithm and establish their convergence rates
in terms of the per-node communications $\mathcal{K}$ and the per-node gradient
evaluations $k$. Our first method, Distributed Nesterov Gradient, achieves
rates $O\left({\log \mathcal{K}}/{\mathcal{K}}\right)$ and $O\left({\log
k}/{k}\right)$. Our second method, Distributed Nesterov gradient with Consensus
iterations, assumes at all nodes knowledge of $L$ and $\mu(W)$ -- the second
largest singular value of the $N \times N$ doubly stochastic weight matrix $W$.
It achieves rates $O\left({1}/{\mathcal{K}^{2-\xi}}\right)$ and
$O\left({1}/{k^2}\right)$ ($\xi>0$ arbitrarily small). Further, we give with
both methods explicit dependence of the convergence constants on $N$ and $W$.
Simulation examples illustrate our findings. |
---|---|
AbstractList | We study distributed optimization problems when $N$ nodes minimize the sum of
their individual costs subject to a common vector variable. The costs are
convex, have Lipschitz continuous gradient (with constant $L$), and bounded
gradient. We propose two fast distributed gradient algorithms based on the
centralized Nesterov gradient algorithm and establish their convergence rates
in terms of the per-node communications $\mathcal{K}$ and the per-node gradient
evaluations $k$. Our first method, Distributed Nesterov Gradient, achieves
rates $O\left({\log \mathcal{K}}/{\mathcal{K}}\right)$ and $O\left({\log
k}/{k}\right)$. Our second method, Distributed Nesterov gradient with Consensus
iterations, assumes at all nodes knowledge of $L$ and $\mu(W)$ -- the second
largest singular value of the $N \times N$ doubly stochastic weight matrix $W$.
It achieves rates $O\left({1}/{\mathcal{K}^{2-\xi}}\right)$ and
$O\left({1}/{k^2}\right)$ ($\xi>0$ arbitrarily small). Further, we give with
both methods explicit dependence of the convergence constants on $N$ and $W$.
Simulation examples illustrate our findings. |
Author | Moura, Jose M. F Jakovetic, Dusan Xavier, Joao |
Author_xml | – sequence: 1 givenname: Dusan surname: Jakovetic fullname: Jakovetic, Dusan – sequence: 2 givenname: Joao surname: Xavier fullname: Xavier, Joao – sequence: 3 givenname: Jose M. F surname: Moura fullname: Moura, Jose M. F |
BackLink | https://doi.org/10.48550/arXiv.1112.2972$$DView paper in arXiv |
BookMark | eNrjYmDJy89LZWCQMDTQM7EwNTXQTyyqyCzTMzQ0NNIzsjQ34mRQdEssLlFwySwuKcpMKi1JTVFwL0pMyUzNK1HwTS3JyE8p5mFgTUvMKU7lhdLcDHJuriHOHrpgs-ILijJzE4sq40FmxoPMNCaoAACzzyw- |
ContentType | Journal Article |
Copyright | http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
Copyright_xml | – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
DBID | AKY AKZ GOX |
DOI | 10.48550/arxiv.1112.2972 |
DatabaseName | arXiv Computer Science arXiv Mathematics arXiv.org |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: GOX name: arXiv.org url: http://arxiv.org/find sourceTypes: Open Access Repository |
DeliveryMethod | fulltext_linktorsrc |
ExternalDocumentID | 1112_2972 |
GroupedDBID | AKY AKZ GOX |
ID | FETCH-arxiv_primary_1112_29723 |
IEDL.DBID | GOX |
IngestDate | Tue Jul 22 23:03:25 EDT 2025 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-arxiv_primary_1112_29723 |
OpenAccessLink | https://arxiv.org/abs/1112.2972 |
ParticipantIDs | arxiv_primary_1112_2972 |
PublicationCentury | 2000 |
PublicationDate | 2011-12-13 |
PublicationDateYYYYMMDD | 2011-12-13 |
PublicationDate_xml | – month: 12 year: 2011 text: 2011-12-13 day: 13 |
PublicationDecade | 2010 |
PublicationYear | 2011 |
Score | 2.9473293 |
SecondaryResourceType | preprint |
Snippet | We study distributed optimization problems when $N$ nodes minimize the sum of
their individual costs subject to a common vector variable. The costs are
convex,... |
SourceID | arxiv |
SourceType | Open Access Repository |
SubjectTerms | Computer Science - Information Theory Mathematics - Information Theory |
Title | Fast Distributed Gradient Methods |
URI | https://arxiv.org/abs/1112.2972 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwY2BQMTQ3TDM1NTbVBcZoiq4J6LChpESjRN2U1CRLE1NTYGkIvhvQ18_MI9TEK8I0golBHrYXJrGoIrMMcj5wUrE-MCMa6RlZmgPLWGYjI1Dfyt0_AjLZCD6JC6ocrgzYwgSLIFURboIM_NC2nYIjJDKEGJhS80QYFN0Si0sUXEDn04KulkpNUXAvAi-zKlHwBd_eXCzKIOfmGuLsoQs2NL4Acv4DqNFsFA-yzliMgQXYSU-VYFBITU4zS7FIMTVPs7QwSTUxS0o2NQQWg6kGqSnAFoaZmSSDOA5DpHDKSDNwgUcwDY10DY1lGFhKikpTZYFVYEmSHDggAHWCX1M |
linkProvider | Cornell University |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Fast+Distributed+Gradient+Methods&rft.au=Jakovetic%2C+Dusan&rft.au=Xavier%2C+Joao&rft.au=Moura%2C+Jose+M.+F&rft.date=2011-12-13&rft_id=info:doi/10.48550%2Farxiv.1112.2972&rft.externalDocID=1112_2972 |