VLIS: Unimodal Language Models Guide Multimodal Language Generation

15 Oct 2023  ยท  Jiwan Chung, Youngjae Yu ยท

Multimodal language generation, which leverages the synergy of language and vision, is a rapidly expanding field. However, existing vision-language models face challenges in tasks that require complex linguistic understanding. To address this issue, we introduce Visual-Language models as Importance Sampling weights (VLIS), a novel framework that combines the visual conditioning capability of vision-language models with the language understanding of unimodal text-only language models without further training. It extracts pointwise mutual information of each image and text from a visual-language model and uses the value as an importance sampling weight to adjust the token likelihood from a text-only model. VLIS improves vision-language models on diverse tasks, including commonsense understanding (WHOOPS, OK-VQA, and ScienceQA) and complex text generation (Concadia, Image Paragraph Captioning, and ROCStories). Our results suggest that VLIS represents a promising new direction for multimodal language generation.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Caption Generation Concadia VLIS (BLIP-2) CIDEr 44.1 # 1
Zero-Shot Image Paragraph Captioning Image Paragraph Captioning VLIS (BLIP-2) METEOR 14.6 # 1
CIDEr 14.8 # 1
BLEU-4 6.4 # 1
Zero-Shot Image Paragraph Captioning Image Paragraph Captioning BLIP-2 METEOR 10.8 # 2
CIDEr 6.5 # 2
BLEU-4 4.9 # 2
Explanation Generation WHOOPS! A Vision-and-Language Benchmark of Synthetic and Compositional Images VLIS (Lynx) Accuracy 80 # 1
Explanation Generation WHOOPS! A Vision-and-Language Benchmark of Synthetic and Compositional Images VLIS (LLaVA) Accuracy 73 # 2

Methods


No methods listed for this paper. Add relevant methods here