Learning a mapping function from an input source video to an output video.
( Image credit: vid2vid )
We study the problem of video-to-video synthesis, whose goal is to learn a mapping function from an input source video (e. g., a sequence of semantic segmentation masks) to an output photorealistic video that precisely depicts the content of the source video.
SEMANTIC SEGMENTATION VIDEO PREDICTION VIDEO-TO-VIDEO SYNTHESIS
To address the limitations, we propose a few-shot vid2vid framework, which learns to synthesize videos of previously unseen subjects or scenes by leveraging few example images of the target at test time.
Ranked #1 on
Video-to-Video Synthesis
on YouTube Dancing
Video inpainting aims to fill spatio-temporal holes with plausible content in a video.
IMAGE INPAINTING OPTICAL FLOW ESTIMATION VIDEO DENOISING VIDEO INPAINTING VIDEO-TO-VIDEO SYNTHESIS
Blind video decaptioning is a problem of automatically removing text overlays and inpainting the occluded parts in videos without any input masks.
We do hope that this series will provide you a big overview of the field, so that you will not need to read all the literature by yourself, independent of your background on GANs.
CONDITIONAL IMAGE GENERATION IMAGE-TO-IMAGE TRANSLATION VIDEO GENERATION VIDEO-TO-VIDEO SYNTHESIS
Videos of actions are complex signals, containing rich compositional structure.