Learning Robust Representations via Multi-View Information Bottleneck

The information bottleneck principle provides an information-theoretic method for representation learning, by training an encoder to retain all information which is relevant for predicting the label while minimizing the amount of other, excess information in the representation. The original formulat...

Full description

Saved in:
Bibliographic Details
Main Authors Federici, Marco, Dutta, Anjan, Forré, Patrick, Kushman, Nate, Akata, Zeynep
Format Journal Article
LanguageEnglish
Published 17.02.2020
Subjects
Online AccessGet full text
DOI10.48550/arxiv.2002.07017

Cover

Loading…
Abstract The information bottleneck principle provides an information-theoretic method for representation learning, by training an encoder to retain all information which is relevant for predicting the label while minimizing the amount of other, excess information in the representation. The original formulation, however, requires labeled data to identify the superfluous information. In this work, we extend this ability to the multi-view unsupervised setting, where two views of the same underlying entity are provided but the label is unknown. This enables us to identify superfluous information as that not shared by both views. A theoretical analysis leads to the definition of a new multi-view model that produces state-of-the-art results on the Sketchy dataset and label-limited versions of the MIR-Flickr dataset. We also extend our theory to the single-view setting by taking advantage of standard data augmentation techniques, empirically showing better generalization capabilities when compared to common unsupervised approaches for representation learning.
AbstractList The information bottleneck principle provides an information-theoretic method for representation learning, by training an encoder to retain all information which is relevant for predicting the label while minimizing the amount of other, excess information in the representation. The original formulation, however, requires labeled data to identify the superfluous information. In this work, we extend this ability to the multi-view unsupervised setting, where two views of the same underlying entity are provided but the label is unknown. This enables us to identify superfluous information as that not shared by both views. A theoretical analysis leads to the definition of a new multi-view model that produces state-of-the-art results on the Sketchy dataset and label-limited versions of the MIR-Flickr dataset. We also extend our theory to the single-view setting by taking advantage of standard data augmentation techniques, empirically showing better generalization capabilities when compared to common unsupervised approaches for representation learning.
Author Dutta, Anjan
Kushman, Nate
Forré, Patrick
Akata, Zeynep
Federici, Marco
Author_xml – sequence: 1
  givenname: Marco
  surname: Federici
  fullname: Federici, Marco
– sequence: 2
  givenname: Anjan
  surname: Dutta
  fullname: Dutta, Anjan
– sequence: 3
  givenname: Patrick
  surname: Forré
  fullname: Forré, Patrick
– sequence: 4
  givenname: Nate
  surname: Kushman
  fullname: Kushman, Nate
– sequence: 5
  givenname: Zeynep
  surname: Akata
  fullname: Akata, Zeynep
BackLink https://doi.org/10.48550/arXiv.2002.07017$$DView paper in arXiv
BookMark eNrjYmDJy89LZWCQNDTQM7EwNTXQTyyqyCzTMzIwMNIzMDcwNOdkcPVJTSzKy8xLVwjKTyotLlEISi0oSi1OzStJLMnMzytWKMtMVPAtzSnJ1A3LTC1X8MxLyy_KBcspOOWXlOSk5qUmZ_MwsKYl5hSn8kJpbgZ5N9cQZw9dsIXxBUWZuYlFlfEgi-PBFhsTVgEAsfM6yQ
ContentType Journal Article
Copyright http://arxiv.org/licenses/nonexclusive-distrib/1.0
Copyright_xml – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0
DBID AKY
EPD
GOX
DOI 10.48550/arxiv.2002.07017
DatabaseName arXiv Computer Science
arXiv Statistics
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 2002_07017
GroupedDBID AKY
EPD
GOX
ID FETCH-arxiv_primary_2002_070173
IEDL.DBID GOX
IngestDate Wed Jul 23 02:00:55 EDT 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-arxiv_primary_2002_070173
OpenAccessLink https://arxiv.org/abs/2002.07017
ParticipantIDs arxiv_primary_2002_07017
PublicationCentury 2000
PublicationDate 2020-02-17
PublicationDateYYYYMMDD 2020-02-17
PublicationDate_xml – month: 02
  year: 2020
  text: 2020-02-17
  day: 17
PublicationDecade 2020
PublicationYear 2020
Score 3.4324377
SecondaryResourceType preprint
Snippet The information bottleneck principle provides an information-theoretic method for representation learning, by training an encoder to retain all information...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Learning
Statistics - Machine Learning
Title Learning Robust Representations via Multi-View Information Bottleneck
URI https://arxiv.org/abs/2002.07017
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV09T8MwED21nVgQCFD59sBqSO04SUdALRUDSBWgbJUvPqMIKUVNWvj52E4QLF19J9vyh-5Ofu8Z4MrqSJBEy-U4KXisUXHMlOHSKkXaYhZRQFs8JbPX-DFXeQ_YLxdGr77LTasPjPWNRxBcu0M5SvvQF8JDth6e8_ZxMkhxdf5_fi7HDE3_gsR0D3a77I7dttuxDz2qDmDSaZi-s_kS13XD5gF92pF-qpptSs0CEZa_lfTFOoaQt7G7pdcYrqj4OITL6eTlfsbDwIvPViXC_-8oFmFO8ggGrpanIbB0hO4WUDw2ZGKjM41SoivkksymuojMMQy39XKy3XQKO8KXgf6fkvQMBs1qTecuVjZ4ERbsB32qb6Y
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Learning+Robust+Representations+via+Multi-View+Information+Bottleneck&rft.au=Federici%2C+Marco&rft.au=Dutta%2C+Anjan&rft.au=Forr%C3%A9%2C+Patrick&rft.au=Kushman%2C+Nate&rft.date=2020-02-17&rft_id=info:doi/10.48550%2Farxiv.2002.07017&rft.externalDocID=2002_07017