Sequential Ensembling for Semantic Segmentation

8 Oct 2022  ·  Rawal Khirodkar, Brandon Smith, Siddhartha Chandra, Amit Agrawal, Antonio Criminisi ·

Ensemble approaches for deep-learning-based semantic segmentation remain insufficiently explored despite the proliferation of competitive benchmarks and downstream applications. In this work, we explore and benchmark the popular ensembling approach of combining predictions of multiple, independently-trained, state-of-the-art models at test time on popular datasets. Furthermore, we propose a novel method inspired by boosting to sequentially ensemble networks that significantly outperforms the naive ensemble baseline. Our approach trains a cascade of models conditioned on class probabilities predicted by the previous model as an additional input. A key benefit of this approach is that it allows for dynamic computation offloading, which helps deploy models on mobile devices. Our proposed novel ADaptive modulatiON (ADON) block allows spatial feature modulation at various layers using previous-stage probabilities. Our approach does not require sophisticated sample selection strategies during training and works with multiple neural architectures. We significantly improve over the naive ensemble baseline on challenging datasets such as Cityscapes, ADE-20K, COCO-Stuff, and PASCAL-Context and set a new state-of-the-art.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Semantic Segmentation ADE20K Sequential Ensemble (SegFormer) Validation mIoU 54 # 63
Params (M) 216.3 # 18
Semantic Segmentation ADE20K Sequential Ensemble (DeepLabv3+) Validation mIoU 46.8 # 162
Semantic Segmentation Cityscapes val Sequential Ensemble (MiT-B5 + HRNet) mIoU 84.8 # 11
Semantic Segmentation PASCAL Context Sequential Ensemble (Segformer + HRNet) mIoU 62.1 # 10

Methods