Image Captioning

613 papers with code • 32 benchmarks • 64 datasets

Image Captioning is the task of describing the content of an image in words. This task lies at the intersection of computer vision and natural language processing. Most image captioning systems use an encoder-decoder framework, where an input image is encoded into an intermediate representation of the information in the image, and then decoded into a descriptive text sequence. The most popular benchmarks are nocaps and COCO, and models are typically evaluated according to a BLEU or CIDER metric.

( Image credit: Reflective Decoding Network for Image Captioning, ICCV'19)

Libraries

Use these libraries to find Image Captioning models and implementations
4 papers
8,685
3 papers
2,319
See all 8 libraries.

Latest papers with no code

On Speculative Decoding for Multimodal Large Language Models

no code yet • 13 Apr 2024

We show that a language-only model can serve as a good draft model for speculative decoding with LLaVA 7B, bypassing the need for image tokens and their associated processing components from the draft model.

View Selection for 3D Captioning via Diffusion Ranking

no code yet • 11 Apr 2024

Scalable annotation approaches are crucial for constructing extensive 3D-text datasets, facilitating a broader range of applications.

Panoptic Perception: A Novel Task and Fine-grained Dataset for Universal Remote Sensing Image Interpretation

no code yet • 6 Apr 2024

Experimental results on FineGrip demonstrate the feasibility of the panoptic perception task and the beneficial effect of multi-task joint optimization on individual tasks.

Would Deep Generative Models Amplify Bias in Future Models?

no code yet • 4 Apr 2024

We investigate the impact of deep generative models on potential social biases in upcoming computer vision models.

Harnessing the Power of Large Vision Language Models for Synthetic Image Detection

no code yet • 3 Apr 2024

This study contributes to the advancement of synthetic image detection by exploiting the capabilities of visual language models such as BLIP-2 and ViTGPT2.

Bi-LORA: A Vision-Language Approach for Synthetic Image Detection

no code yet • 2 Apr 2024

Advancements in deep image synthesis techniques, such as generative adversarial networks (GANs) and diffusion models (DMs), have ushered in an era of generating highly realistic images.

VLRM: Vision-Language Models act as Reward Models for Image Captioning

no code yet • 2 Apr 2024

In this work, we present an unsupervised method for enhancing an image captioning model (in our case, BLIP2) using reinforcement learning and vision-language models like CLIP and BLIP2-ITM as reward models.

LLaMA-Excitor: General Instruction Tuning via Indirect Feature Interaction

no code yet • 1 Apr 2024

LLaMA-Excitor ensures a self-adaptive allocation of additional attention to input instructions, thus effectively preserving LLMs' pre-trained knowledge when fine-tuning LLMs on low-quality instruction-following datasets.

Learning by Correction: Efficient Tuning Task for Zero-Shot Generative Vision-Language Reasoning

no code yet • 1 Apr 2024

Generative vision-language models (VLMs) have shown impressive performance in zero-shot vision-language tasks like image captioning and visual question answering.

LocCa: Visual Pretraining with Location-aware Captioners

no code yet • 28 Mar 2024

In this paper, we propose a simple visual pretraining method with location-aware captioners (LocCa).