Deep Ranking with Adaptive Margin Triplet Loss
We propose a simple modification from a fixed margin triplet loss to an adaptive margin triplet loss. While the original triplet loss is used widely in classification problems such as face recognition, face re-identification and fine-grained similarity, our proposed loss is well suited for rating da...
Saved in:
Main Authors | , |
---|---|
Format | Journal Article |
Language | English |
Published |
13.07.2021
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | We propose a simple modification from a fixed margin triplet loss to an
adaptive margin triplet loss. While the original triplet loss is used widely in
classification problems such as face recognition, face re-identification and
fine-grained similarity, our proposed loss is well suited for rating datasets
in which the ratings are continuous values. In contrast to original triplet
loss where we have to sample data carefully, in out method, we can generate
triplets using the whole dataset, and the optimization can still converge
without frequently running into a model collapsing issue. The adaptive margins
only need to be computed once before the training, which is much less expensive
than generating triplets after every epoch as in the fixed margin case. Besides
substantially improved training stability (the proposed model never collapsed
in our experiments compared to a couple of times that the training collapsed on
existing triplet loss), we achieved slightly better performance than the
original triplet loss on various rating datasets and network architectures. |
---|---|
DOI: | 10.48550/arxiv.2107.06187 |