SoftTriple Loss: Deep Metric Learning Without Triplet Sampling
Distance metric learning (DML) is to learn the embeddings where examples from the same class are closer than examples from different classes. It can be cast as an optimization problem with triplet constraints. Due to the vast number of triplet constraints, a sampling strategy is essential for DML. W...
Main Authors: | , , , , , |
---|---|
Format: | Article in Journal/Newspaper |
Language: | unknown |
Published: |
arXiv
2019
|
Subjects: | |
Online Access: | https://dx.doi.org/10.48550/arxiv.1909.05235 https://arxiv.org/abs/1909.05235 |
id |
ftdatacite:10.48550/arxiv.1909.05235 |
---|---|
record_format |
openpolar |
spelling |
ftdatacite:10.48550/arxiv.1909.05235 2023-05-15T16:01:28+02:00 SoftTriple Loss: Deep Metric Learning Without Triplet Sampling Qian, Qi Shang, Lei Sun, Baigui Hu, Juhua Li, Hao Jin, Rong 2019 https://dx.doi.org/10.48550/arxiv.1909.05235 https://arxiv.org/abs/1909.05235 unknown arXiv arXiv.org perpetual, non-exclusive license http://arxiv.org/licenses/nonexclusive-distrib/1.0/ Computer Vision and Pattern Recognition cs.CV FOS Computer and information sciences Article CreativeWork article Preprint 2019 ftdatacite https://doi.org/10.48550/arxiv.1909.05235 2022-03-10T16:41:39Z Distance metric learning (DML) is to learn the embeddings where examples from the same class are closer than examples from different classes. It can be cast as an optimization problem with triplet constraints. Due to the vast number of triplet constraints, a sampling strategy is essential for DML. With the tremendous success of deep learning in classifications, it has been applied for DML. When learning embeddings with deep neural networks (DNNs), only a mini-batch of data is available at each iteration. The set of triplet constraints has to be sampled within the mini-batch. Since a mini-batch cannot capture the neighbors in the original set well, it makes the learned embeddings sub-optimal. On the contrary, optimizing SoftMax loss, which is a classification loss, with DNN shows a superior performance in certain DML tasks. It inspires us to investigate the formulation of SoftMax. Our analysis shows that SoftMax loss is equivalent to a smoothed triplet loss where each class has a single center. In real-world data, one class can contain several local clusters rather than a single one, e.g., birds of different poses. Therefore, we propose the SoftTriple loss to extend the SoftMax loss with multiple centers for each class. Compared with conventional deep metric learning algorithms, optimizing SoftTriple loss can learn the embeddings without the sampling phase by mildly increasing the size of the last fully connected layer. Experiments on the benchmark fine-grained data sets demonstrate the effectiveness of the proposed loss function. Code is available at https://github.com/idstcv/SoftTriple : accepted by ICCV'19 Article in Journal/Newspaper DML DataCite Metadata Store (German National Library of Science and Technology) |
institution |
Open Polar |
collection |
DataCite Metadata Store (German National Library of Science and Technology) |
op_collection_id |
ftdatacite |
language |
unknown |
topic |
Computer Vision and Pattern Recognition cs.CV FOS Computer and information sciences |
spellingShingle |
Computer Vision and Pattern Recognition cs.CV FOS Computer and information sciences Qian, Qi Shang, Lei Sun, Baigui Hu, Juhua Li, Hao Jin, Rong SoftTriple Loss: Deep Metric Learning Without Triplet Sampling |
topic_facet |
Computer Vision and Pattern Recognition cs.CV FOS Computer and information sciences |
description |
Distance metric learning (DML) is to learn the embeddings where examples from the same class are closer than examples from different classes. It can be cast as an optimization problem with triplet constraints. Due to the vast number of triplet constraints, a sampling strategy is essential for DML. With the tremendous success of deep learning in classifications, it has been applied for DML. When learning embeddings with deep neural networks (DNNs), only a mini-batch of data is available at each iteration. The set of triplet constraints has to be sampled within the mini-batch. Since a mini-batch cannot capture the neighbors in the original set well, it makes the learned embeddings sub-optimal. On the contrary, optimizing SoftMax loss, which is a classification loss, with DNN shows a superior performance in certain DML tasks. It inspires us to investigate the formulation of SoftMax. Our analysis shows that SoftMax loss is equivalent to a smoothed triplet loss where each class has a single center. In real-world data, one class can contain several local clusters rather than a single one, e.g., birds of different poses. Therefore, we propose the SoftTriple loss to extend the SoftMax loss with multiple centers for each class. Compared with conventional deep metric learning algorithms, optimizing SoftTriple loss can learn the embeddings without the sampling phase by mildly increasing the size of the last fully connected layer. Experiments on the benchmark fine-grained data sets demonstrate the effectiveness of the proposed loss function. Code is available at https://github.com/idstcv/SoftTriple : accepted by ICCV'19 |
format |
Article in Journal/Newspaper |
author |
Qian, Qi Shang, Lei Sun, Baigui Hu, Juhua Li, Hao Jin, Rong |
author_facet |
Qian, Qi Shang, Lei Sun, Baigui Hu, Juhua Li, Hao Jin, Rong |
author_sort |
Qian, Qi |
title |
SoftTriple Loss: Deep Metric Learning Without Triplet Sampling |
title_short |
SoftTriple Loss: Deep Metric Learning Without Triplet Sampling |
title_full |
SoftTriple Loss: Deep Metric Learning Without Triplet Sampling |
title_fullStr |
SoftTriple Loss: Deep Metric Learning Without Triplet Sampling |
title_full_unstemmed |
SoftTriple Loss: Deep Metric Learning Without Triplet Sampling |
title_sort |
softtriple loss: deep metric learning without triplet sampling |
publisher |
arXiv |
publishDate |
2019 |
url |
https://dx.doi.org/10.48550/arxiv.1909.05235 https://arxiv.org/abs/1909.05235 |
genre |
DML |
genre_facet |
DML |
op_rights |
arXiv.org perpetual, non-exclusive license http://arxiv.org/licenses/nonexclusive-distrib/1.0/ |
op_doi |
https://doi.org/10.48550/arxiv.1909.05235 |
_version_ |
1766397301735030784 |