GMRNet: A Novel Geometric Mean Relation Network for Few-Shot Very Similar Object Classification

With the widespread use of Deep Learning (DL), the use of DL has increased to provide a solution to the problem of object recognition and classification. In addition to classifying many different types of objects, the Deep Metrics Learning(DML) technique is effective in classifying objects that are...

Full description

Bibliographic Details
Published in:IEEE Access
Main Authors: Canan Tastimur, Erhan Akin
Format: Article in Journal/Newspaper
Language:English
Published: IEEE 2022
Subjects:
DML
Online Access:https://doi.org/10.1109/ACCESS.2022.3206528
https://doaj.org/article/3a11628824d24c2ebed3b906a035662a
Description
Summary:With the widespread use of Deep Learning (DL), the use of DL has increased to provide a solution to the problem of object recognition and classification. In addition to classifying many different types of objects, the Deep Metrics Learning(DML) technique is effective in classifying objects that are visually very similar to each other. In this study, a novel Relation Network (RN) based DML has been designed to classify objects in two different datasets we created. We distinguished groups of objects that had a high degree of similarity to each other. These objects have been categorized using few-shot learning(FSL) since they are quite similar to one another. The impact of changing the number of classes and samples in the database on the network’s performance has been studied. It is shown how the network’s accuracy varies depending on the N-way (number of classes) and K-shots (number of samples) combinations used in its design. Additionally, the performance of the network has improved by an average of 15% thanks to the contribution of the recently introduced geometric mean module to the RN in our study. The accuracy rate of our recommended RN in screw and spare parts datasets is 96.1% and 92.3%, respectively. The first dataset consists of 1800 screw images with 18 classes, while the second dataset consists of 4100 spare parts images with 20 classes. The effectiveness of our method is expressed by the two datasets that we have extensively experimentally studied.