While the community has seen many advances in recent years to address the challenging problem of Finegrained Visual Categorization (FGVC), progress seems to be slowing—new state-of-the-art methods often distinguish themselves by improving top-1 accuracy by mere tenths of a percent. However, across all of the now-standard FGVC datasets, there remain sizeable portions of the test data that none of the current state-of-the-art (SOTA) models can successfully predict. This paper provides a framework for identifying and studying the errors that current methods make across diverse fine-grained datasets. Three models of difficulty—Prediction Overlap, Prediction Rank and Pairwise Class Confusion—are employed to highlight the most challenging sets of images and classes. Extensive experiments apply a range of standard and SOTA methods, evaluating them on multiple FGVC domains and datasets. Insights acquired from coupling these difficulty paradigms with the careful analysis of experimental results suggest crucial areas for future FGVC research, focusing critically on the set of elusive images that none of the current models can correctly classify.
We propose three methods for analyzing image and class difficulty: prediction overlap, prediction rank, and pairwise class confusion. We use these methods to analyze difficulty in FGVC datasets.
We propose the iCub dataset as an additional large-scale source of validation images for testing models trained to recognize birds from the 200 species found in CUB-200. iCub images are sourced from iNaturalist, and have a different visual distribution. This allows us to test CUB models on a more difficult set of data to see how they generalize.
For our analysis, we train five models (using different random seeds) for each of the model types. Accuracy of these models is shown below. We find that the models tend to underperform the results reported in their respective papers when they are reimplemented in a standardized context. We found that WS-DAN performed particularly well, despite being an older approach.
We also show results of applying our proposed difficulty analysis methods: prediction overlap, prediction rank, and similar class confusion.
Finally, we show some examples of difficult images. We notice some patterns in these images such as camoflauge, occlusion, non-standard poses, or multiple objects. We hypothesize that these factors contribute to the image difficulty; but we must be careful not to confuse correlation with causation. Determining the real or exact reasons for misclassification is a challenging task.
@inproceedings{anderson2024elusive-images,
title={Elusive Images: Beyond Coarse Analysis for Fine-grained Recognition},
author={Anderson, Connor and Gwilliam, Matt and Gaskin, Evelyn and Farrell, Ryan},
booktitle={WACV},
year={2024}
}