Bridging the Gap between Object and Image-level Representations for Open-Vocabulary Detection

Existing open-vocabulary object detectors typically enlarge their vocabulary sizes by leveraging different forms of weak supervision. This helps generalize to novel objects at inference. Two popular forms of weak-supervision used in open-vocabulary detection (OVD) include pretrained CLIP model and image-level supervision. We note that both these modes of supervision are not optimally aligned for the detection task: CLIP is trained with image-text pairs and lacks precise localization of objects while the image-level supervision has been used with heuristics that do not accurately specify local object regions. In this work, we propose to address this problem by performing object-centric alignment of the language embeddings from the CLIP model. Furthermore, we visually ground the objects with only image-level supervision using a pseudo-labeling process that provides high-quality object proposals and helps expand the vocabulary during training. We establish a bridge between the above two object-alignment strategies via a novel weight transfer function that aggregates their complimentary strengths. In essence, the proposed model seeks to minimize the gap between object and image-centric representations in the OVD setting. On the COCO benchmark, our proposed approach achieves 36.6 AP50 on novel classes, an absolute 8.2 gain over the previous best performance. For LVIS, we surpass the state-of-the-art ViLD model by 5.0 mask AP for rare categories and 3.4 overall. Code: https://github.com/hanoonaR/object-centric-ovd.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Open Vocabulary Object Detection LVIS v1.0 Object-Centric-OVD AP novel-LVIS base training 21.1 # 14
Zero-Shot Object Detection MSCOCO Object-Centric-OVD AP 0.5 40.5 # 2
Open Vocabulary Object Detection MSCOCO Object-Centric-OVD AP 0.5 36.9 # 9
Open Vocabulary Object Detection Objects365 Object-Centric-OVD mask AP50 22.3 # 1
Open Vocabulary Object Detection OpenImages-v4 Object-Centric-OVD mask AP50 42.9 # 1
Open Vocabulary Attribute Detection OVAD benchmark Object-Centric-OVD (ResNet50) mean average precision 14.6 # 5

Methods