A Simple Framework for Open-Vocabulary Segmentation and Detection

We present OpenSeeD, a simple Open-vocabulary Segmentation and Detection framework that jointly learns from different segmentation and detection datasets. To bridge the gap of vocabulary and annotation granularity, we first introduce a pre-trained text encoder to encode all the visual concepts in two tasks and learn a common semantic space for them. This gives us reasonably good results compared with the counterparts trained on segmentation task only. To further reconcile them, we locate two discrepancies: $i$) task discrepancy -- segmentation requires extracting masks for both foreground objects and background stuff, while detection merely cares about the former; $ii$) data discrepancy -- box and mask annotations are with different spatial granularity, and thus not directly interchangeable. To address these issues, we propose a decoupled decoding to reduce the interference between foreground/background and a conditioned mask decoding to assist in generating masks for given boxes. To this end, we develop a simple encoder-decoder model encompassing all three techniques and train it jointly on COCO and Objects365. After pre-training, our model exhibits competitive or stronger zero-shot transferability for both segmentation and detection. Specifically, OpenSeeD beats the state-of-the-art method for open-vocabulary instance and panoptic segmentation across 5 datasets, and outperforms previous work for open-vocabulary detection on LVIS and ODinW under similar settings. When transferred to specific tasks, our model achieves new SoTA for panoptic segmentation on COCO and ADE20K, and instance segmentation on ADE20K and Cityscapes. Finally, we note that OpenSeeD is the first to explore the potential of joint training on segmentation and detection, and hope it can be received as a strong baseline for developing a single model for both tasks in open world.

PDF Abstract ICCV 2023 PDF ICCV 2023 Abstract

Results from the Paper


Ranked #2 on Instance Segmentation on ADE20K val (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Benchmark
Instance Segmentation ADE20K val OpenSeeD AP 42.6 # 2
Panoptic Segmentation ADE20K val OpenSeed(SwinL, single scale, 1280x1280) PQ 53.7 # 2
Instance Segmentation Cityscapes val OpenSeeD( SwinL, single-scale) mask AP 49.3 # 2
Panoptic Segmentation COCO minival OpenSeeD (SwinL, single-scale) PQ 59.5 # 2
AP 53.2 # 1
Zero Shot Segmentation Segmentation in the Wild OpenSEED Mean AP 36.1 # 7

Methods


No methods listed for this paper. Add relevant methods here