24 papers with code • 0 benchmarks • 1 datasets
We propose a novel end-to-end deep neural network that generates dynamic upsampling filters and a residual image, which are computed depending on the local spatio-temporal neighborhood of each pixel to avoid explicit motion compensation.
Ranked #2 on Video Super-Resolution on Vid4 - 4x upscaling
In this paper, we propose a deformable 3D convolution network (D3Dnet) to incorporate spatio-temporal information from both spatial and temporal dimensions for video SR.
Extensive experiments demonstrate that HR optical flows provide more accurate correspondences than their LR counterparts and improve both accuracy and consistency performance.
Ranked #9 on Video Super-Resolution on Vid4 - 4x upscaling
Most previous fusion strategies either fail to fully utilize temporal information or cost too much time, and how to effectively fuse temporal information from consecutive frames plays an important role in video super-resolution (SR).
In this work, we propose a motion estimation and motion compensation driven neural network for video frame interpolation.
Ranked #5 on Video Frame Interpolation on Middlebury