Global Convergence of Sub-gradient Method for Robust Matrix Recovery: Small Initialization, Noisy Measurements, and Over-parameterization

In this work, we study the performance of sub-gradient method (SubGM) on a natural nonconvex and nonsmooth formulation of low-rank matrix recovery with $\ell_1$-loss, where the goal is to recover a low-rank matrix from a limited number of measurements, a subset of which may be grossly corrupted with...

Full description

Saved in:
Bibliographic Details
Main Authors Ma, Jianhao, Fattahi, Salar
Format Journal Article
LanguageEnglish
Published 17.02.2022
Subjects
Online AccessGet full text

Cover

Loading…
Abstract In this work, we study the performance of sub-gradient method (SubGM) on a natural nonconvex and nonsmooth formulation of low-rank matrix recovery with $\ell_1$-loss, where the goal is to recover a low-rank matrix from a limited number of measurements, a subset of which may be grossly corrupted with noise. We study a scenario where the rank of the true solution is unknown and over-estimated instead. The over-estimation of the rank gives rise to an over-parameterized model in which there are more degrees of freedom than needed. Such over-parameterization may lead to overfitting, or adversely affect the performance of the algorithm. We prove that a simple SubGM with small initialization is agnostic to both over-parameterization and noise in the measurements. In particular, we show that small initialization nullifies the effect of over-parameterization on the performance of SubGM, leading to an exponential improvement in its convergence rate. Moreover, we provide the first unifying framework for analyzing the behavior of SubGM under both outlier and Gaussian noise models, showing that SubGM converges to the true solution, even under arbitrarily large and arbitrarily dense noise values, and--perhaps surprisingly--even if the globally optimal solutions do not correspond to the ground truth. At the core of our results is a robust variant of restricted isometry property, called Sign-RIP, which controls the deviation of the sub-differential of the $\ell_1$-loss from that of an ideal, expected loss. As a byproduct of our results, we consider a subclass of robust low-rank matrix recovery with Gaussian measurements, and show that the number of required samples to guarantee the global convergence of SubGM is independent of the over-parameterized rank.
AbstractList In this work, we study the performance of sub-gradient method (SubGM) on a natural nonconvex and nonsmooth formulation of low-rank matrix recovery with $\ell_1$-loss, where the goal is to recover a low-rank matrix from a limited number of measurements, a subset of which may be grossly corrupted with noise. We study a scenario where the rank of the true solution is unknown and over-estimated instead. The over-estimation of the rank gives rise to an over-parameterized model in which there are more degrees of freedom than needed. Such over-parameterization may lead to overfitting, or adversely affect the performance of the algorithm. We prove that a simple SubGM with small initialization is agnostic to both over-parameterization and noise in the measurements. In particular, we show that small initialization nullifies the effect of over-parameterization on the performance of SubGM, leading to an exponential improvement in its convergence rate. Moreover, we provide the first unifying framework for analyzing the behavior of SubGM under both outlier and Gaussian noise models, showing that SubGM converges to the true solution, even under arbitrarily large and arbitrarily dense noise values, and--perhaps surprisingly--even if the globally optimal solutions do not correspond to the ground truth. At the core of our results is a robust variant of restricted isometry property, called Sign-RIP, which controls the deviation of the sub-differential of the $\ell_1$-loss from that of an ideal, expected loss. As a byproduct of our results, we consider a subclass of robust low-rank matrix recovery with Gaussian measurements, and show that the number of required samples to guarantee the global convergence of SubGM is independent of the over-parameterized rank.
Author Ma, Jianhao
Fattahi, Salar
Author_xml – sequence: 1
  givenname: Jianhao
  surname: Ma
  fullname: Ma, Jianhao
– sequence: 2
  givenname: Salar
  surname: Fattahi
  fullname: Fattahi, Salar
BackLink https://doi.org/10.48550/arXiv.2202.08788$$DView paper in arXiv
BookMark eNotkM1OAjEUhbvQhaIP4Mo-AIOX6XR-3BmiSIKSAHty295ik5mWdAYCvoFv7YisTnJyvrP4btmVD54YexjDKCulhCeMR3cYpSmkIyiLsrxhP9M6KKz5JPgDxS15TTxYvtqrZBvROPId_6DuKxhuQ-TLoPZt32AX3ZEvSYeeOj3zVYN1zWfedQ5r942dC37IP4NrTz2O7T5S01-1Q47e8EUPJTuM2FBH8TK_Y9cW65buLzlg67fX9eQ9mS-ms8nLPMG8KBNjBQijcwtZWYlUZhmM0eRGgkZpCTSkUKnCoraAAFJliqpCVIXKtShyKQbs8f_27GKzi67BeNr8OdmcnYhfVctgyA
ContentType Journal Article
Copyright http://creativecommons.org/licenses/by/4.0
Copyright_xml – notice: http://creativecommons.org/licenses/by/4.0
DBID AKY
AKZ
EPD
GOX
DOI 10.48550/arxiv.2202.08788
DatabaseName arXiv Computer Science
arXiv Mathematics
arXiv Statistics
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 2202_08788
GroupedDBID AKY
AKZ
EPD
GOX
ID FETCH-LOGICAL-a678-df303dc6f04893254401ad6d50ca5fe0c0209b7facf0a005b4be97397b6c37653
IEDL.DBID GOX
IngestDate Mon Jan 08 05:43:56 EST 2024
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-a678-df303dc6f04893254401ad6d50ca5fe0c0209b7facf0a005b4be97397b6c37653
OpenAccessLink https://arxiv.org/abs/2202.08788
ParticipantIDs arxiv_primary_2202_08788
PublicationCentury 2000
PublicationDate 2022-02-17
PublicationDateYYYYMMDD 2022-02-17
PublicationDate_xml – month: 02
  year: 2022
  text: 2022-02-17
  day: 17
PublicationDecade 2020
PublicationYear 2022
Score 1.8337361
SecondaryResourceType preprint
Snippet In this work, we study the performance of sub-gradient method (SubGM) on a natural nonconvex and nonsmooth formulation of low-rank matrix recovery with...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Learning
Mathematics - Optimization and Control
Statistics - Machine Learning
Title Global Convergence of Sub-gradient Method for Robust Matrix Recovery: Small Initialization, Noisy Measurements, and Over-parameterization
URI https://arxiv.org/abs/2202.08788
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV1NTwIxEG2AkxejUYOfmYNHGpfCdsGbISKaAIliwo10tltDAruGBQM_wX_tTHf9uHjtzlymaefN9s0bIa6xS7WOS0KpHRrZbnasRK2Y4WhbVnVijf71fDjSg9f20zScVgR898KY1Xb-UegDY36jFOtpdqhMq4qqUkzZehhPi8dJL8VV2v_aEcb0S3-SRP9A7JfoDu6K7TgUlSQ9Ep-Frj70mODtex0TyBzQkZVvK8-4WsPQD3IGQpDwnOEmpxXWzt8Cl4fktbuFl6VZLOCRyT5mUXZPNmCUzfMduf_86csbYFILY3KSrOu9ZL5LaX4sJv37SW8gywkI0lASkdZRgrGxdgFLxLCYWNA0VtswiE3okiAmrNfFyJnYBYaOE7Yx6UaEMFDHdHGErRNRS7M0qQvAZtRSlIdiy5iBUBFiGDmFTkXkHOCpqPu4zd4LkYsZh3TmQ3r2_6dzsae4HYAHpEQXorZebZJLStJrvPI79QXCtpR9
link.rule.ids 228,230,783,888
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Global+Convergence+of+Sub-gradient+Method+for+Robust+Matrix+Recovery%3A+Small+Initialization%2C+Noisy+Measurements%2C+and+Over-parameterization&rft.au=Ma%2C+Jianhao&rft.au=Fattahi%2C+Salar&rft.date=2022-02-17&rft_id=info:doi/10.48550%2Farxiv.2202.08788&rft.externalDocID=2202_08788