Comparison of distance metric learning methods against label noise for fine-grained recognition

https://doi.org/10.1117/12.2587246(2021)

引用 0|浏览0
暂无评分
摘要
Deep metric learning is an approach to establish a distance metric between data to measure the similarity. Metric learning methods map an input image to a representative feature space where semantically similar samples are closer and dissimilar examples are far away. The use of metric learning in fine-grained recognition is widely studied in recent years. Fine-grained Recognition (FGR) focuses on categorizing hard-to-distinguish classes such as birds' species and models of the cars. In the FGR datasets, the intra-class variance is high, while the inter-class variance is low. This makes it challenging to annotate them, leading to erroneous labels. Especially in defense applications, it is quite costly to label the data due to the fact that this work should be done by experts. The performance of the metric learning methods is directly related to the loss function exploited during training the model. The loss functions are divided into two categories: Pair-based and proxy-based approaches. A proxy is a representation of distribution in feature space. Although the pair-based loss functions utilize the data-to-data relations, the proxy-based loss functions exploit data-to-proxy relations. In this paper, we analyzed the effect of the label noise on open-set fine-grained recognition performance. The pair-based and proxy-based methods are evaluated on three widely adopted benchmark datasets: CUB 200-2011, Stanford Cars 196, and FGVC Aircraft.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要