Depth-Aware Video Frame Interpolation

Video frame interpolation aims to synthesize nonexistent frames in-between the original frames. While significant advances have been made from the recent deep convolutional neural networks, the quality of interpolation is often reduced due to large object motion or occlusion. In this work, we propose a video frame interpolation method which explicitly detects the occlusion by exploring the depth information. Specifically, we develop a depth-aware flow projection layer to synthesize intermediate flows that preferably sample closer objects than farther ones. In addition, we learn hierarchical features to gather contextual information from neighboring pixels. The proposed model then warps the input frames, depth maps, and contextual features based on the optical flow and local interpolation kernels for synthesizing the output frame. Our model is compact, efficient, and fully differentiable. Quantitative and qualitative results demonstrate that the proposed model performs favorably against state-of-the-art frame interpolation methods on a wide variety of datasets.

PDF Abstract CVPR 2019 PDF CVPR 2019 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Video Frame Interpolation Middlebury DAIN Interpolation Error 4.86 # 5
Video Frame Interpolation UCF101 DAIN PSNR 34.99 # 15
SSIM 0.9683 # 12
Video Frame Interpolation Vimeo90K DAIN PSNR 34.71 # 18
PSNR 34.71 # 18
SSIM 0.9756 # 13
Video Frame Interpolation X4K1000FPS DAIN_f PSNR 27.52 # 12
SSIM 0.821 # 11
tOF 3.47 # 3
Video Frame Interpolation X4K1000FPS DAIN PSNR 26.78 # 13
SSIM 0.807 # 12
tOF 3.83 # 4

Methods


No methods listed for this paper. Add relevant methods here