PolyMaX: General Dense Prediction with Mask Transformer

Dense prediction tasks, such as semantic segmentation, depth estimation, and surface normal prediction, can be easily formulated as per-pixel classification (discrete outputs) or regression (continuous outputs). This per-pixel prediction paradigm has remained popular due to the prevalence of fully convolutional networks. However, on the recent frontier of segmentation task, the community has been witnessing a shift of paradigm from per-pixel prediction to cluster-prediction with the emergence of transformer architectures, particularly the mask transformers, which directly predicts a label for a mask instead of a pixel. Despite this shift, methods based on the per-pixel prediction paradigm still dominate the benchmarks on the other dense prediction tasks that require continuous outputs, such as depth estimation and surface normal prediction. Motivated by the success of DORN and AdaBins in depth estimation, achieved by discretizing the continuous output space, we propose to generalize the cluster-prediction based method to general dense prediction tasks. This allows us to unify dense prediction tasks with the mask transformer framework. Remarkably, the resulting model PolyMaX demonstrates state-of-the-art performance on three benchmarks of NYUD-v2 dataset. We hope our simple yet effective design can inspire more research on exploiting mask transformers for more dense prediction tasks. Code and model will be made available.

PDF Abstract

Datasets


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Surface Normals Estimation NYU Depth v2 PolyMaX(ConvNeXt-L) % < 11.25 65.66 # 2
% < 22.5 82.28 # 2
% < 30 87.83 # 2
Mean Angle Error 13.09 # 2
RMSE 20.4 # 3
Semantic Segmentation NYU Depth v2 PolyMaX(ConvNeXt-L) Mean IoU 58.08% # 5
Monocular Depth Estimation NYU-Depth V2 PolyMaX(ConvNeXt-L) RMSE 0.25 # 11
absolute relative error 0.067 # 11
Delta < 1.25 0.969 # 9
Delta < 1.25^2 0.9958 # 10
Delta < 1.25^3 0.999 # 4
log 10 0.029 # 10

Methods


No methods listed for this paper. Add relevant methods here