Image Classification
3781 papers with code • 168 benchmarks • 240 datasets
Image Classification is a fundamental task in vision recognition that aims to understand and categorize an image as a whole under a specific label. Unlike object detection, which involves classification and location of multiple objects within an image, image classification typically pertains to single-object images. When the classification becomes highly detailed or reaches instance-level, it is often referred to as image retrieval, which also involves finding similar images in a large database.
Libraries
Use these libraries to find Image Classification models and implementationsDatasets
Subtasks
- Out of Distribution (OOD) Detection
- Few-Shot Image Classification
- Fine-Grained Image Classification
- Semi-Supervised Image Classification
- Semi-Supervised Image Classification
- Learning with noisy labels
- Hyperspectral Image Classification
- Self-Supervised Image Classification
- Small Data Image Classification
- Multi-Label Image Classification
- Genre classification
- Sequential Image Classification
- Unsupervised Image Classification
- Efficient ViTs
- Document Image Classification
- Satellite Image Classification
- Sparse Representation-based Classification
- Photo geolocation estimation
- Image Classification with Differential Privacy
- Token Reduction
- Superpixel Image Classification
- Classification Consistency
- Gallbladder Cancer Detection
- Artistic style classification
- Artist classification
- Temporal Metadata Manipulation Detection
- Misclassification Rate - Natural Adversarial Samples
- Scale Generalisation
Latest papers with no code
Brain Storm Optimization Based Swarm Learning for Diabetic Retinopathy Image Classification
Federated learning, as a privacy-preserving machine learning architecture, has shown promising performance in balancing data privacy and model utility by keeping private data on the client's side and using a central server to coordinate a set of clients for model training through aggregating their uploaded model parameters.
A review of deep learning-based information fusion techniques for multimodal medical image classification
Multimodal medical imaging plays a pivotal role in clinical diagnosis and research, as it combines information from various imaging modalities to provide a more comprehensive understanding of the underlying pathology.
An MRP Formulation for Supervised Learning: Generalized Temporal Difference Learning Models
In traditional statistical learning, data points are usually assumed to be independently and identically distributed (i. i. d.)
Deep multi-prototype capsule networks
Capsule networks are a type of neural network that identify image parts and form the instantiation parameters of a whole hierarchically.
CKD: Contrastive Knowledge Distillation from A Sample-wise Perspective
Note that constraints on intra-sample similarities and inter-sample dissimilarities can be efficiently and effectively reformulated into a contrastive learning framework with newly designed positive and negative pairs.
WangLab at MEDIQA-M3G 2024: Multimodal Medical Answer Generation using Large Language Models
This paper outlines our submission to the MEDIQA2024 Multilingual and Multimodal Medical Answer Generation (M3G) shared task.
EncodeNet: A Framework for Boosting DNN Accuracy with Entropy-driven Generalized Converting Autoencoder
Our experimental results demonstrate that EncodeNet improves the accuracy of VGG16 from 92. 64% to 94. 05% on CIFAR-10 and RestNet20 from 74. 56% to 76. 04% on CIFAR-100.
I2CANSAY:Inter-Class Analogical Augmentation and Intra-Class Significance Analysis for Non-Exemplar Online Task-Free Continual Learning
Online task-free continual learning (OTFCL) is a more challenging variant of continual learning which emphasizes the gradual shift of task boundaries and learns in an online mode.
Nested-TNT: Hierarchical Vision Transformers with Multi-Scale Feature Processing
ViT divides an image into several local patches, known as "visual sentences".
Transformer-Based Classification Outcome Prediction for Multimodal Stroke Treatment
This study proposes a multi-modal fusion framework Multitrans based on the Transformer architecture and self-attention mechanism.