Ranked List Loss for Deep Metric Learning

The objective of deep metric learning (DML) is to learn embeddings that can capture semantic similarity and dissimilarity information among data points. Existing pairwise or tripletwise loss functions used in DML are known to suffer from slow convergence due to a large proportion of trivial pairs or...

Full description

Bibliographic Details
Main Authors: Wang, Xinshao, Hua, Yang, Kodirov, Elyor, Robertson, Neil M.
Format: Report
Language:unknown
Published: arXiv 2019
Subjects:
DML
Online Access:https://dx.doi.org/10.48550/arxiv.1903.03238
https://arxiv.org/abs/1903.03238
id ftdatacite:10.48550/arxiv.1903.03238
record_format openpolar
spelling ftdatacite:10.48550/arxiv.1903.03238 2023-05-15T16:01:42+02:00 Ranked List Loss for Deep Metric Learning Wang, Xinshao Hua, Yang Kodirov, Elyor Robertson, Neil M. 2019 https://dx.doi.org/10.48550/arxiv.1903.03238 https://arxiv.org/abs/1903.03238 unknown arXiv arXiv.org perpetual, non-exclusive license http://arxiv.org/licenses/nonexclusive-distrib/1.0/ Computer Vision and Pattern Recognition cs.CV FOS Computer and information sciences Preprint Article article CreativeWork 2019 ftdatacite https://doi.org/10.48550/arxiv.1903.03238 2022-04-01T08:35:56Z The objective of deep metric learning (DML) is to learn embeddings that can capture semantic similarity and dissimilarity information among data points. Existing pairwise or tripletwise loss functions used in DML are known to suffer from slow convergence due to a large proportion of trivial pairs or triplets as the model improves. To improve this, ranking-motivated structured losses are proposed recently to incorporate multiple examples and exploit the structured information among them. They converge faster and achieve state-of-the-art performance. In this work, we unveil two limitations of existing ranking-motivated structured losses and propose a novel ranked list loss to solve both of them. First, given a query, only a fraction of data points is incorporated to build the similarity structure. Consequently, some useful examples are ignored and the structure is less informative. To address this, we propose to build a set-based similarity structure by exploiting all instances in the gallery. The learning setting can be interpreted as few-shot retrieval: given a mini-batch, every example is iteratively used as a query, and the rest ones compose the gallery to search, i.e., the support set in few-shot setting. The rest examples are split into a positive set and a negative set. For every mini-batch, the learning objective of ranked list loss is to make the query closer to the positive set than to the negative set by a margin. Second, previous methods aim to pull positive pairs as close as possible in the embedding space. As a result, the intraclass data distribution tends to be extremely compressed. In contrast, we propose to learn a hypersphere for each class in order to preserve useful similarity structure inside it, which functions as regularisation. Extensive experiments demonstrate the superiority of our proposal by comparing with the state-of-the-art methods. : Accepted to T-PAMI. Therefore, to read the offical version, please go to IEEE Xplore. Fine-grained image retrieval task. Our source code is available online: https://github.com/XinshaoAmosWang/Ranked-List-Loss-for-DML Report DML DataCite Metadata Store (German National Library of Science and Technology) The Gallery ENVELOPE(-86.417,-86.417,72.535,72.535) Triplets ENVELOPE(-59.750,-59.750,-62.383,-62.383)
institution Open Polar
collection DataCite Metadata Store (German National Library of Science and Technology)
op_collection_id ftdatacite
language unknown
topic Computer Vision and Pattern Recognition cs.CV
FOS Computer and information sciences
spellingShingle Computer Vision and Pattern Recognition cs.CV
FOS Computer and information sciences
Wang, Xinshao
Hua, Yang
Kodirov, Elyor
Robertson, Neil M.
Ranked List Loss for Deep Metric Learning
topic_facet Computer Vision and Pattern Recognition cs.CV
FOS Computer and information sciences
description The objective of deep metric learning (DML) is to learn embeddings that can capture semantic similarity and dissimilarity information among data points. Existing pairwise or tripletwise loss functions used in DML are known to suffer from slow convergence due to a large proportion of trivial pairs or triplets as the model improves. To improve this, ranking-motivated structured losses are proposed recently to incorporate multiple examples and exploit the structured information among them. They converge faster and achieve state-of-the-art performance. In this work, we unveil two limitations of existing ranking-motivated structured losses and propose a novel ranked list loss to solve both of them. First, given a query, only a fraction of data points is incorporated to build the similarity structure. Consequently, some useful examples are ignored and the structure is less informative. To address this, we propose to build a set-based similarity structure by exploiting all instances in the gallery. The learning setting can be interpreted as few-shot retrieval: given a mini-batch, every example is iteratively used as a query, and the rest ones compose the gallery to search, i.e., the support set in few-shot setting. The rest examples are split into a positive set and a negative set. For every mini-batch, the learning objective of ranked list loss is to make the query closer to the positive set than to the negative set by a margin. Second, previous methods aim to pull positive pairs as close as possible in the embedding space. As a result, the intraclass data distribution tends to be extremely compressed. In contrast, we propose to learn a hypersphere for each class in order to preserve useful similarity structure inside it, which functions as regularisation. Extensive experiments demonstrate the superiority of our proposal by comparing with the state-of-the-art methods. : Accepted to T-PAMI. Therefore, to read the offical version, please go to IEEE Xplore. Fine-grained image retrieval task. Our source code is available online: https://github.com/XinshaoAmosWang/Ranked-List-Loss-for-DML
format Report
author Wang, Xinshao
Hua, Yang
Kodirov, Elyor
Robertson, Neil M.
author_facet Wang, Xinshao
Hua, Yang
Kodirov, Elyor
Robertson, Neil M.
author_sort Wang, Xinshao
title Ranked List Loss for Deep Metric Learning
title_short Ranked List Loss for Deep Metric Learning
title_full Ranked List Loss for Deep Metric Learning
title_fullStr Ranked List Loss for Deep Metric Learning
title_full_unstemmed Ranked List Loss for Deep Metric Learning
title_sort ranked list loss for deep metric learning
publisher arXiv
publishDate 2019
url https://dx.doi.org/10.48550/arxiv.1903.03238
https://arxiv.org/abs/1903.03238
long_lat ENVELOPE(-86.417,-86.417,72.535,72.535)
ENVELOPE(-59.750,-59.750,-62.383,-62.383)
geographic The Gallery
Triplets
geographic_facet The Gallery
Triplets
genre DML
genre_facet DML
op_rights arXiv.org perpetual, non-exclusive license
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
op_doi https://doi.org/10.48550/arxiv.1903.03238
_version_ 1766397451182276608