Supervised Contrastive Learning with Multiple Positive Examples
The present disclosure provides an improved training methodology that enables supervised contrastive learning to be simultaneously performed across multiple positive and negative training examples. In particular, example aspects of the present disclosure are directed to an improved, supervised versi...
Saved in:
Main Authors | , , , , , , , , |
---|---|
Format | Patent |
Language | English |
Published |
18.05.2023
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | The present disclosure provides an improved training methodology that enables supervised contrastive learning to be simultaneously performed across multiple positive and negative training examples. In particular, example aspects of the present disclosure are directed to an improved, supervised version of the batch contrastive loss, which has been shown to be very effective at learning powerful representations in the self-supervised setting Thus, the proposed techniques adapt contrastive learning to the fully supervised setting and also enable learning to occur simultaneously across multiple positive examples. |
---|---|
Bibliography: | Application Number: US202117920623 |