Softmax Splatting for Video Frame Interpolation

CVPR 2020  ·  Simon Niklaus, Feng Liu ·

Differentiable image sampling in the form of backward warping has seen broad adoption in tasks like depth estimation and optical flow prediction. In contrast, how to perform forward warping has seen less attention, partly due to additional challenges such as resolving the conflict of mapping multiple pixels to the same target location in a differentiable way. We propose softmax splatting to address this paradigm shift and show its effectiveness on the application of frame interpolation. Specifically, given two input frames, we forward-warp the frames and their feature pyramid representations based on an optical flow estimate using softmax splatting. In doing so, the softmax splatting seamlessly handles cases where multiple source pixels map to the same target location. We then use a synthesis network to predict the interpolation result from the warped representations. Our softmax splatting allows us to not only interpolate frames at an arbitrary time but also to fine tune the feature pyramid and the optical flow. We show that our synthesis approach, empowered by softmax splatting, achieves new state-of-the-art results for video frame interpolation.

PDF Abstract CVPR 2020 PDF CVPR 2020 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Video Frame Interpolation Middlebury SoftSplat Interpolation Error 4.223 # 2
SSIM 0.971 # 1
PSNR 38.42 # 2
Video Frame Interpolation UCF101 SoftSplat PSNR 35.39 # 7
SSIM 0.952 # 13
Video Frame Interpolation Vimeo90K SoftSplat PSNR 36.10 # 8
SSIM 0.970 # 15

Methods