Towards a Hypothesis on Visual Transformation based Self-Supervision

We propose the first qualitative hypothesis characterizing the behavior of visual transformation based self-supervision, called the VTSS hypothesis. Given a dataset upon which a self-supervised task is performed while predicting instantiations of a transformation, the hypothesis states that if the p...

Full description

Saved in:
Bibliographic Details
Main Authors Pal, Dipan K, Nallamothu, Sreena, Savvides, Marios
Format Journal Article
LanguageEnglish
Published 24.11.2019
Subjects
Online AccessGet full text

Cover

Loading…
Abstract We propose the first qualitative hypothesis characterizing the behavior of visual transformation based self-supervision, called the VTSS hypothesis. Given a dataset upon which a self-supervised task is performed while predicting instantiations of a transformation, the hypothesis states that if the predicted instantiations of the transformations are already present in the dataset, then the representation learned will be less useful. The hypothesis was derived by observing a key constraint in the application of self-supervision using a particular transformation. This constraint, which we term the transformation conflict for this paper, forces a network learn degenerative features thereby reducing the usefulness of the representation. The VTSS hypothesis helps us identify transformations that have the potential to be effective as a self-supervision task. Further, it helps to generally predict whether a particular transformation based self-supervision technique would be effective or not for a particular dataset. We provide extensive evaluations on CIFAR 10, CIFAR 100, SVHN and FMNIST confirming the hypothesis and the trends it predicts. We also propose novel cost-effective self-supervision techniques based on translation and scale, which when combined with rotation outperforms all transformations applied individually. Overall, this paper aims to shed light on the phenomenon of visual transformation based self-supervision.
AbstractList We propose the first qualitative hypothesis characterizing the behavior of visual transformation based self-supervision, called the VTSS hypothesis. Given a dataset upon which a self-supervised task is performed while predicting instantiations of a transformation, the hypothesis states that if the predicted instantiations of the transformations are already present in the dataset, then the representation learned will be less useful. The hypothesis was derived by observing a key constraint in the application of self-supervision using a particular transformation. This constraint, which we term the transformation conflict for this paper, forces a network learn degenerative features thereby reducing the usefulness of the representation. The VTSS hypothesis helps us identify transformations that have the potential to be effective as a self-supervision task. Further, it helps to generally predict whether a particular transformation based self-supervision technique would be effective or not for a particular dataset. We provide extensive evaluations on CIFAR 10, CIFAR 100, SVHN and FMNIST confirming the hypothesis and the trends it predicts. We also propose novel cost-effective self-supervision techniques based on translation and scale, which when combined with rotation outperforms all transformations applied individually. Overall, this paper aims to shed light on the phenomenon of visual transformation based self-supervision.
Author Pal, Dipan K
Nallamothu, Sreena
Savvides, Marios
Author_xml – sequence: 1
  givenname: Dipan K
  surname: Pal
  fullname: Pal, Dipan K
– sequence: 2
  givenname: Sreena
  surname: Nallamothu
  fullname: Nallamothu, Sreena
– sequence: 3
  givenname: Marios
  surname: Savvides
  fullname: Savvides, Marios
BackLink https://doi.org/10.48550/arXiv.1911.10594$$DView paper in arXiv
BookMark eNotj71OwzAUhT3AAIUHYMIvkOAb22k8ovJTpEoMjVij6_haWErjyG4LfXtCYTpH33B0vmt2McaRGLsDUapGa_GA6TscSzAAJQht1BV7auMXJpc58vVpivtPyiHzOPKPkA848DbhmH1MO9yHmVrM5PiWBl9sDxOlY8gzvmGXHodMt_-5YO3Lc7taF5v317fV46bAeqkKI62xrq4q7UiYSjQgsRcNeU1WQD8XqPqlnX-CdNYbbwih1rpXSpL1KBfs_m_2rNFNKewwnbpfne6sI38AUQdH6Q
ContentType Journal Article
Copyright http://arxiv.org/licenses/nonexclusive-distrib/1.0
Copyright_xml – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0
DBID AKY
EPD
GOX
DOI 10.48550/arxiv.1911.10594
DatabaseName arXiv Computer Science
arXiv Statistics
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 1911_10594
GroupedDBID AKY
EPD
GOX
ID FETCH-LOGICAL-a674-93b9bd6225de0920813ac08ef5eb01c8ef12c7b55013dbf9f9ea1655c443ebfa3
IEDL.DBID GOX
IngestDate Mon Jan 08 05:39:08 EST 2024
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-a674-93b9bd6225de0920813ac08ef5eb01c8ef12c7b55013dbf9f9ea1655c443ebfa3
OpenAccessLink https://arxiv.org/abs/1911.10594
ParticipantIDs arxiv_primary_1911_10594
PublicationCentury 2000
PublicationDate 2019-11-24
PublicationDateYYYYMMDD 2019-11-24
PublicationDate_xml – month: 11
  year: 2019
  text: 2019-11-24
  day: 24
PublicationDecade 2010
PublicationYear 2019
Score 1.7559358
SecondaryResourceType preprint
Snippet We propose the first qualitative hypothesis characterizing the behavior of visual transformation based self-supervision, called the VTSS hypothesis. Given a...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Computer Vision and Pattern Recognition
Computer Science - Learning
Statistics - Machine Learning
Title Towards a Hypothesis on Visual Transformation based Self-Supervision
URI https://arxiv.org/abs/1911.10594
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV1LT8MwDLa2nbggEKDxVA5cK_pIuuaIgFEhAYcV1FsVN4lUCY1pXSf49zhpEbtwixxfbMv67MQPgOuEQEnFEgPMkAfcpiZQKRkkS1FqioCFrF2D8_NLmr_xp1KUI2C_vTBq_dVs-_nA2N5QMhG5TbSSj2Ecx65k6_G17D8n_Siugf-Pj2JMT9oBifkB7A_RHbvtzXEII7M8gvvCl6a2TLH8e-U6ntqmZZ9L9t60HXEXO7EjUR2uaLYwHzZYdCvnyu5B6xiK-UNxlwfD8gKSdcYDmaBEnZK3aBPKmIA3UXWYGSsMhlFNhyiuZ0giRIlGK600KkqFqDlPDFqVnMCE8n8zBRbyOhRWCOQEt2gipBwDNXeDy6xAhacw9SJXq34-ReW0UXltnP1_dQ57hP3StdXF_AImm3VnLglfN3jllfwDQCF7mg
link.rule.ids 228,230,783,888
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Towards+a+Hypothesis+on+Visual+Transformation+based+Self-Supervision&rft.au=Pal%2C+Dipan+K&rft.au=Nallamothu%2C+Sreena&rft.au=Savvides%2C+Marios&rft.date=2019-11-24&rft_id=info:doi/10.48550%2Farxiv.1911.10594&rft.externalDocID=1911_10594