MoVie: Revisiting Modulated Convolutions for Visual Counting and Beyond

ICLR 2021  ·  Duy-Kien Nguyen, Vedanuj Goswami, Xinlei Chen ·

This paper focuses on visual counting, which aims to predict the number of occurrences given a natural image and a query (e.g. a question or a category). Unlike most prior works that use explicit, symbolic models which can be computationally expensive and limited in generalization, we propose a simple and effective alternative by revisiting modulated convolutions that fuse the query and the image locally. Following the design of residual bottleneck, we call our method MoVie, short for Modulated conVolutional bottlenecks. Notably, MoVie reasons implicitly and holistically and only needs a single forward-pass during inference. Nevertheless, MoVie showcases strong performance for counting: 1) advancing the state-of-the-art on counting-specific VQA tasks while being more efficient; 2) outperforming prior-art on difficult benchmarks like COCO for common object counting; 3) helped us secure the first place of 2020 VQA challenge when integrated as a module for 'number' related questions in generic VQA models. Finally, we show evidence that modulated convolutions such as MoVie can serve as a general mechanism for reasoning tasks beyond counting.

PDF Abstract ICLR 2021 PDF ICLR 2021 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Object Counting HowMany-QA MoVie-ResNeXt Accuracy 64 # 1
RMSE 2.3 # 1
Object Counting HowMany-QA MoVie Accuracy 61.2 # 2
RMSE 2.36 # 3
Object Counting TallyQA-Complex MoVie-ResNeXt Accuracy 56.8 # 4
RMSE 1.43 # 1
Object Counting TallyQA-Complex MoVie Accuracy 54.1 # 6
RMSE 1.52 # 3
Object Counting TallyQA-Simple MoVie-ResNeXt Accuracy 74.9 # 4
RMSE 1 # 1
Object Counting TallyQA-Simple MoVie Accuracy 70.8 # 6
RMSE 1.09 # 2

Methods


No methods listed for this paper. Add relevant methods here