1 code implementation • ECCV 2020 • Jan Brejcha, Michal Lukáč, Yannick Hold-Geoffroy, Oliver Wang, Martin Čadík
We introduce a solution to large scale Augmented Reality for outdoor scenes by registering camera images to textured Digital Elevation Models (DEMs).
Ranked #2 on
Patch Matching
on HPatches
(using extra training data)
no code implementations • 15 Jul 2023 • Jiahui Huang, Leonid Sigal, Kwang Moo Yi, Oliver Wang, Joon-Young Lee
We present Interactive Neural Video Editing (INVE), a real-time video editing solution, which can assist the video editing process by consistently propagating sparse frame edits to the entire video clip.
no code implementations • CVPR 2023 • Linyi Jin, Jianming Zhang, Yannick Hold-Geoffroy, Oliver Wang, Kevin Matzen, Matthew Sticha, David F. Fouhey
We propose perspective fields as a representation that models the local perspective properties of an image.
no code implementations • 27 Nov 2022 • Animesh Karnewar, Oliver Wang, Tobias Ritschel, Niloy Mitra
We introduce 3inGAN, an unconditional 3D generative model trained from 2D images of a single self-similar 3D scene.
no code implementations • 22 Nov 2022 • David Chuan-En Lin, Fabian Caba Heilbron, Joon-Young Lee, Oliver Wang, Nikolas Martelaro
This paper investigates the challenge of extracting highlight moments from videos.
no code implementations • 22 Nov 2022 • David Chuan-En Lin, Fabian Caba Heilbron, Joon-Young Lee, Oliver Wang, Nikolas Martelaro
Video has become a dominant form of media.
1 code implementation • 28 Aug 2022 • Wei Yin, Jianming Zhang, Oliver Wang, Simon Niklaus, Simon Chen, Yifan Liu, Chunhua Shen
To do so, we propose a two-stage framework that first predicts depth up to an unknown scale and shift from a single monocular image, and then exploits 3D point cloud data to predict the depth shift and the camera's focal length that allow us to recover 3D scene shapes.
1 code implementation • 29 Jul 2022 • Wei Yin, Jianming Zhang, Oliver Wang, Simon Niklaus, Simon Chen, Chunhua Shen
Our method leverages a data driven prior in the form of a single image depth prediction network trained on large-scale datasets, the output of which is used as an input to our model.
no code implementations • CVPR 2022 • Zhongzheng Ren, Aseem Agarwala, Bryan Russell, Alexander G. Schwing, Oliver Wang
We introduce an approach for selecting objects in neural volumetric 3D representations, such as multi-plane images (MPI) and neural radiance fields (NeRF).
no code implementations • 22 May 2022 • Animesh Karnewar, Tobias Ritschel, Oliver Wang, Niloy J. Mitra
Although the MLPs are able to represent complex scenes with unprecedented quality and memory footprint, this expressive power of the MLPs, however, comes at the cost of long training and inference times.
2 code implementations • 23 Sep 2021 • Yoni Kasten, Dolev Ofri, Oliver Wang, Tali Dekel
We present a method that decomposes, or "unwraps", an input video into a set of layered 2D atlases, each providing a unified representation of the appearance of an object (or background) over the video.
no code implementations • 24 Jun 2021 • Youssef A. Mejjati, Isa Milefchik, Aaron Gokaslan, Oliver Wang, Kwang In Kim, James Tompkin
We present an algorithm that learns a coarse 3D representation of objects from unposed multi-view 2D mask supervision, then uses it to generate detailed mask and image texture.
2 code implementations • 11 May 2021 • Marco A. Martínez Ramírez, Oliver Wang, Paris Smaragdis, Nicholas J. Bryan
We present a data-driven approach to automate audio signal processing by incorporating stateful third-party, audio effects as layers within a deep neural network.
no code implementations • 5 Feb 2021 • Felix Klose, Oliver Wang, Jean-Charles Bazin, Marcus Magnor, Alexander Sorkine-Hornung
We present a novel, sampling-based framework for processing video that enables high-quality scene-space video effects in the presence of inevitable errors in depth and camera pose estimation.
1 code implementation • 5 Feb 2021 • Tobias Hinz, Matthew Fisher, Oliver Wang, Eli Shechtman, Stefan Wermter
Our model generates novel poses based on keypoint locations, which can be modified in real time while providing interactive feedback, allowing for intuitive reposing and animation.
no code implementations • 12 Jan 2021 • Jan Rueegg, Oliver Wang, Aljoscha Smolic, Markus Gross
DuctTake is a system designed to enable practical compositing of multiple takes of a scene into a single video.
1 code implementation • CVPR 2021 • Wei Yin, Jianming Zhang, Oliver Wang, Simon Niklaus, Long Mai, Simon Chen, Chunhua Shen
Despite significant progress in monocular depth estimation in the wild, recent state-of-the-art methods cannot be used to recover accurate 3D scene shape due to an unknown depth shift induced by shift-invariant reconstruction losses used in mixed-data depth prediction training, and possible unknown camera focal length.
Ranked #1 on
Indoor Monocular Depth Estimation
on DIODE
(using extra training data)
3 code implementations • CVPR 2021 • Zhengqi Li, Simon Niklaus, Noah Snavely, Oliver Wang
We present a method to perform novel view and time synthesis of dynamic scenes, requiring only a monocular video with known camera poses as input.
no code implementations • 2 Nov 2020 • Simon Niklaus, Long Mai, Oliver Wang
Video frame interpolation, the synthesis of novel views in time, is an increasingly popular research direction with many new papers further advancing the state of the art.
1 code implementation • 7 Jul 2020 • Ping Hu, Federico Perazzi, Fabian Caba Heilbron, Oliver Wang, Zhe Lin, Kate Saenko, Stan Sclaroff
The proposed architecture relies on our fast spatial attention, which is a simple yet efficient modification of the popular self-attention mechanism and captures the same rich spatial context at a small fraction of the computational cost, by changing the order of operations.
Ranked #32 on
Semantic Segmentation
on DensePASS
4 code implementations • NeurIPS 2020 • Taesung Park, Jun-Yan Zhu, Oliver Wang, Jingwan Lu, Eli Shechtman, Alexei A. Efros, Richard Zhang
Deep generative models have become increasingly effective at producing realistic images from randomly sampled seeds, but using such models for controllable manipulation of existing images remains challenging.
1 code implementation • CVPR 2020 • Ping Hu, Fabian Caba Heilbron, Oliver Wang, Zhe Lin, Stan Sclaroff, Federico Perazzi
We present TDNet, a temporally distributed network designed for fast and accurate video semantic segmentation.
Ranked #2 on
Video Semantic Segmentation
on Cityscapes val
3 code implementations • 25 Mar 2020 • Tobias Hinz, Matthew Fisher, Oliver Wang, Stefan Wermter
Recently there has been an interest in the potential of learning generative models from a single image, as opposed to from a large dataset.
1 code implementation • 1 Jan 2020 • Youssef Alami Mejjati, Zejiang Shen, Michael Snower, Aaron Gokaslan, Oliver Wang, James Tompkin, Kwang In Kim
We present an algorithm to generate diverse foreground objects and composite them into background images using a GAN architecture.
4 code implementations • CVPR 2020 • Sheng-Yu Wang, Oliver Wang, Richard Zhang, Andrew Owens, Alexei A. Efros
In this work we ask whether it is possible to create a "universal" detector for telling apart real images from these generated by a CNN, regardless of architecture or dataset used.
1 code implementation • ICCV 2019 • Arnab Ghosh, Richard Zhang, Puneet K. Dokania, Oliver Wang, Alexei A. Efros, Philip H. S. Torr, Eli Shechtman
We propose an interactive GAN-based sketch-to-image translation method that helps novice users create images of simple objects.
2 code implementations • ICCV 2019 • Sheng-Yu Wang, Oliver Wang, Andrew Owens, Richard Zhang, Alexei A. Efros
Most malicious photo manipulations are created using standard image editing tools, such as Adobe Photoshop.
no code implementations • 25 Apr 2019 • Chaoyang Wang, Simon Lucey, Federico Perazzi, Oliver Wang
We present a fully data-driven method to compute depth from diverse monocular video sequences that contain large amounts of non-rigid objects, e. g., people.
1 code implementation • CVPR 2019 • Chen-Hsuan Lin, Oliver Wang, Bryan C. Russell, Eli Shechtman, Vladimir G. Kim, Matthew Fisher, Simon Lucey
In this paper, we address the problem of 3D object mesh reconstruction from RGB videos.
4 code implementations • CVPR 2020 • Animesh Karnewar, Oliver Wang
While Generative Adversarial Networks (GANs) have seen huge successes in image synthesis tasks, they are notoriously difficult to adapt to different datasets, in part due to instability during training and sensitivity to hyperparameters.
Ranked #1 on
Image Generation
on Indian Celebs 256 x 256
no code implementations • 28 Feb 2019 • Bernd Huber, Hijung Valentina Shin, Bryan Russell, Oliver Wang, Gautham J. Mysore
In video production, inserting B-roll is a widely used technique to enrich the story and make a video more engaging.
no code implementations • NeurIPS 2018 • Pedro Morgado, Nuno Nvasconcelos, Timothy Langlois, Oliver Wang
We introduce an approach to convert mono audio recorded by a 360° video camera into spatial audio, a representation of the distribution of sound over the full viewing sphere.
no code implementations • 18 Oct 2018 • Lijun Wang, Xiaohui Shen, Jianming Zhang, Oliver Wang, Zhe Lin, Chih-Yao Hsieh, Sarah Kong, Huchuan Lu
To achieve this, we propose a novel neural network model comprised of a depth prediction module, a lens blur module, and a guided upsampling module.
1 code implementation • 7 Sep 2018 • Pedro Morgado, Nuno Vasconcelos, Timothy Langlois, Oliver Wang
Using our approach, we show that it is possible to infer the spatial location of sound sources based only on 360 video and a mono audio track.
1 code implementation • EMNLP 2018 • Lisa Anne Hendricks, Oliver Wang, Eli Shechtman, Josef Sivic, Trevor Darrell, Bryan Russell
To benchmark whether our model, and other recent video localization models, can effectively reason about temporal language, we collect the novel TEMPOral reasoning in video and language (TEMPO) dataset.
1 code implementation • ECCV 2018 • Wei-Sheng Lai, Jia-Bin Huang, Oliver Wang, Eli Shechtman, Ersin Yumer, Ming-Hsuan Yang
Our method takes the original unprocessed and per-frame processed videos as inputs to produce a temporally consistent video.
2 code implementations • CVPR 2018 • Chen-Hsuan Lin, Ersin Yumer, Oliver Wang, Eli Shechtman, Simon Lucey
We address the problem of finding realistic geometric corrections to a foreground object such that it appears natural when composited into a background image.
24 code implementations • CVPR 2018 • Richard Zhang, Phillip Isola, Alexei A. Efros, Eli Shechtman, Oliver Wang
We systematically evaluate deep features across different architectures and tasks and compare them with classic metrics.
Ranked #19 on
Video Quality Assessment
on MSU FR VQA Database
6 code implementations • NeurIPS 2017 • Jun-Yan Zhu, Richard Zhang, Deepak Pathak, Trevor Darrell, Alexei A. Efros, Oliver Wang, Eli Shechtman
Our proposed method encourages bijective consistency between the latent encoding and output modes.
no code implementations • 16 Aug 2017 • Chengzhou Tang, Oliver Wang, Ping Tan
Many monocular visual SLAM algorithms are derived from incremental structure-from-motion (SfM) methods.
2 code implementations • ICCV 2017 • Lisa Anne Hendricks, Oliver Wang, Eli Shechtman, Josef Sivic, Trevor Darrell, Bryan Russell
A key obstacle to training our MCN model is that current video datasets do not include pairs of localized video segments and referring expressions, or text descriptions which uniquely identify a corresponding moment.
1 code implementation • CVPR 2017 • Shuochen Su, Mauricio Delbracio, Jue Wang, Guillermo Sapiro, Wolfgang Heidrich, Oliver Wang
We show that the features learned from this dataset extend to deblurring motion blur that arises due to camera shake in a wide range of videos, and compare the quality of results to a number of other baselines.
1 code implementation • CVPR 2016 • Patrick Wieschollek, Oliver Wang, Alexander Sorkine-Hornung, Hendrik P. A. Lensch
We present a new approach for efficient approximate nearest neighbor (ANN) search in high dimensional spaces, extending the idea of Product Quantization.
1 code implementation • CVPR 2017 • Chao Yang, Xin Lu, Zhe Lin, Eli Shechtman, Oliver Wang, Hao Li
Recent advances in deep learning have shown exciting promise in filling large holes in natural images with semantically plausible and context aware details, impacting fundamental image manipulation tasks such as object removal.
1 code implementation • 25 Nov 2016 • Shuochen Su, Mauricio Delbracio, Jue Wang, Guillermo Sapiro, Wolfgang Heidrich, Oliver Wang
We show that the features learned from this dataset extend to deblurring motion blur that arises due to camera shake in a wide range of videos, and compare the quality of results to a number of other baselines.
no code implementations • CVPR 2016 • Nicolas Maerki, Federico Perazzi, Oliver Wang, Alexander Sorkine-Hornung
In this work, we propose a novel approach to video segmentation that operates in bilateral space.
Ranked #76 on
Semi-Supervised Video Object Segmentation
on DAVIS 2016
Semi-Supervised Video Object Segmentation
Video Segmentation
+1
no code implementations • ICCV 2015 • Federico Perazzi, Oliver Wang, Markus Gross, Alexander Sorkine-Hornung
We present a novel approach to video segmentation using multiple object proposals.
Ranked #77 on
Semi-Supervised Video Object Segmentation
on DAVIS 2016
Semi-Supervised Video Object Segmentation
Video Segmentation
+1
no code implementations • ICCV 2015 • Charles Malleson, Jean-Charles Bazin, Oliver Wang, Derek Bradley, Thabo Beeler, Adrian Hilton, Alexander Sorkine-Hornung
We present a method to continuously blend between multiple facial performances of an actor, which can contain different facial expressions or emotional states.
no code implementations • CVPR 2015 • Benjamin Resch, Hendrik P. A. Lensch, Oliver Wang, Marc Pollefeys, Alexander Sorkine-Hornung
Videos consisting of thousands of high resolution frames are challenging for existing structure from motion (SfM) and simultaneous-localization and mapping (SLAM) techniques.
no code implementations • CVPR 2015 • Simone Meyer, Oliver Wang, Henning Zimmer, Max Grosse, Alexander Sorkine-Hornung
Standard approaches to computing interpolated (in-between) frames in a video sequence require accurate pixel correspondences between images e. g. using optical flow.