Depthformer : Multiscale Vision Transformer For Monocular Depth Estimation With Local Global Information Fusion

10 Jul 2022  ·  Ashutosh Agarwal, Chetan Arora ·

Attention-based models such as transformers have shown outstanding performance on dense prediction tasks, such as semantic segmentation, owing to their capability of capturing long-range dependency in an image. However, the benefit of transformers for monocular depth prediction has seldom been explored so far. This paper benchmarks various transformer-based models for the depth estimation task on an indoor NYUV2 dataset and an outdoor KITTI dataset. We propose a novel attention-based architecture, Depthformer for monocular depth estimation that uses multi-head self-attention to produce the multiscale feature maps, which are effectively combined by our proposed decoder network. We also propose a Transbins module that divides the depth range into bins whose center value is estimated adaptively per image. The final depth estimated is a linear combination of bin centers for each pixel. Transbins module takes advantage of the global receptive field using the transformer module in the encoding stage. Experimental results on NYUV2 and KITTI depth estimation benchmark demonstrate that our proposed method improves the state-of-the-art by 3.3%, and 3.3% respectively in terms of Root Mean Squared Error (RMSE). Code is available at https://github.com/ashutosh1807/Depthformer.git.

PDF Abstract

Datasets


Results from the Paper


Ranked #29 on Monocular Depth Estimation on KITTI Eigen split (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Monocular Depth Estimation KITTI Eigen split Depthformer absolute relative error 0.058 # 29
RMSE 2.285 # 27
Sq Rel 0.187 # 3
RMSE log 0.087 # 26
Delta < 1.25 0.967 # 26
Delta < 1.25^2 0.996 # 25
Delta < 1.25^3 0.999 # 11
Monocular Depth Estimation NYU-Depth V2 Depthformer RMSE 0.345 # 33
absolute relative error 0.100 # 38
Delta < 1.25 0.913 # 35
Delta < 1.25^2 0.988 # 28
Delta < 1.25^3 0.997 # 27
log 10 0.042 # 33

Methods


No methods listed for this paper. Add relevant methods here