The Fine-Grained Image Classification task focuses on differentiating between hard-to-distinguish object classes, such as species of birds, flowers, or animals; and identifying the makes or models of vehicles.
( Image credit: Looking for the Devil in the Details )
|TREND||DATASET||BEST METHOD||PAPER TITLE||PAPER||CODE||COMPARE|
In our implementation, we have designed a search space where a policy consists of many sub-policies, one of which is randomly chosen for each image in each mini-batch.
Ranked #3 on Image Classification on SVHN
Deep residual nets are foundations of our submissions to ILSVRC & COCO 2015 competitions, where we also won the 1st places on the tasks of ImageNet detection, ImageNet localization, COCO detection, and COCO segmentation.
Ranked #2 on Pedestrian Attribute Recognition on UAV-Human
DOMAIN GENERALIZATION FINE-GRAINED IMAGE CLASSIFICATION IMAGE-TO-IMAGE TRANSLATION OBJECT DETECTION PEDESTRIAN ATTRIBUTE RECOGNITION PEDESTRIAN TRAJECTORY PREDICTION PERSON RE-IDENTIFICATION RETINAL OCT DISEASE CLASSIFICATION SEMANTIC SEGMENTATION
By stacking the TNT blocks, we build the TNT model for image recognition.
While the Transformer architecture has become the de-facto standard for natural language processing tasks, its applications to computer vision remain limited.
Ranked #1 on Fine-Grained Image Classification on Oxford-IIIT Pets (using extra training data)
In this work, we introduce a series of architecture modifications that aim to boost neural networks' accuracy, while retaining their GPU training and inference efficiency.
Ranked #1 on Fine-Grained Image Classification on Stanford Cars (using extra training data)
Convolutional Neural Networks (ConvNets) are commonly developed at a fixed resource budget, and then scaled up for better accuracy if more resources are available.
Ranked #2 on Fine-Grained Image Classification on Birdsnap (using extra training data)
In this work, we produce a competitive convolution-free transformer by training on Imagenet only.
Ranked #2 on Image Classification on iNaturalist 2018
Scaling up deep neural network capacity has been known as an effective approach to improving model quality for several different machine learning tasks.
Ranked #3 on Fine-Grained Image Classification on Birdsnap (using extra training data)
In consideration of intrinsic consistency between informativeness of the regions and their probability being ground-truth class, we design a novel training paradigm, which enables Navigator to detect most informative regions under the guidance from Teacher.
Ranked #16 on Fine-Grained Image Classification on Stanford Cars
Towards addressing this problem, we propose an iterative matrix square root normalization method for fast end-to-end training of global covariance pooling networks.
Ranked #5 on Fine-Grained Image Classification on CUB-200-2011