no code implementations • 2 Jan 2025 • Zhenyu Li, Wenqing Cui, Shariq Farooq Bhat, Peter Wonka
While current high-resolution depth estimation methods achieve strong results, they often suffer from computational inefficiencies due to reliance on heavyweight models and multiple inference steps, increasing inference time.
no code implementations • 9 Dec 2024 • Ziya Erkoç, Can Gümeli, Chaoyang Wang, Matthias Nießner, Angela Dai, Peter Wonka, Hsin-Ying Lee, Peiye Zhuang
The edited 3D mesh aligns well with the prompts, and remains identical for regions that are not intended to be altered.
no code implementations • 9 Dec 2024 • Bingchen Gong, Diego Gomez, Abdullah Hamdi, Abdelrahman Eldesokey, Ahmed Abdelreheem, Peter Wonka, Maks Ovsjanikov
Experimental evaluations demonstrate that our approach achieves competitive performance on standard benchmarks compared to supervised methods, despite not requiring any 3D keypoint annotations during training.
no code implementations • 5 Dec 2024 • Chaoyang Wang, Peiye Zhuang, Tuan Duc Ngo, Willi Menapace, Aliaksandr Siarohin, Michael Vasilkovsky, Ivan Skorokhodov, Sergey Tulyakov, Peter Wonka, Hsin-Ying Lee
We propose a novel two-stream architecture.
no code implementations • 3 Dec 2024 • Zhenyu Li, Mykola Lavreniuk, Jian Shi, Shariq Farooq Bhat, Peter Wonka
Amodal depth estimation aims to predict the depth of occluded (invisible) parts of objects in a scene.
no code implementations • 29 Nov 2024 • Vadim Pryadilshchikov, Alexander Markin, Artem Komarichev, Ruslan Rakhimov, Peter Wonka, Evgeny Burnaev
We propose a novel framework to remove transient objects from input videos for 3D scene reconstruction using Gaussian Splatting.
no code implementations • 25 Nov 2024 • Biao Zhang, Jing Ren, Peter Wonka
Neural representations of 3D data have been widely adopted across various applications, particularly in recent work leveraging coordinate-based networks to model scalar or vector fields.
no code implementations • 21 Nov 2024 • Jian Shi, Qian Wang, Zhenyu Li, Peter Wonka
Generating high-quality stereo videos that mimic human binocular vision requires maintaining consistent depth perception and temporal coherence across frames.
no code implementations • 2 Oct 2024 • Biao Zhang, Peter Wonka
Different from previous approaches that only work on a regular image or volume grid, our hierarchical autoencoder operates on unordered sets of vectors.
1 code implementation • 30 Sep 2024 • Jian Shi, Zhenyu Li, Peter Wonka
We introduce \textit{ImmersePro}, an innovative framework specifically designed to transform single-view videos into stereo videos.
no code implementations • 27 Aug 2024 • Abdelrahman Eldesokey, Peter Wonka
Nevertheless, existing approaches for layout control are limited to 2D layouts, require the user to provide a static layout beforehand, and fail to preserve generated images under layout changes.
no code implementations • 21 Jun 2024 • Savva Ignatyev, Nina Konovalova, Daniil Selikhanovych, Oleg Voynov, Nikolay Patakin, Ilya Olkov, Dmitry Senushkin, Alexey Artemov, Anton Konushin, Alexander Filippov, Peter Wonka, Evgeny Burnaev
In order to achieve the alignment of the corresponding parts of the generated objects, we propose to embed these objects into a common latent space and optimize the continuous transitions between these objects.
no code implementations • 18 Jun 2024 • Jing Gu, Yuwei Fang, Ivan Skorokhodov, Peter Wonka, Xinya Du, Sergey Tulyakov, Xin Eric Wang
Video editing is a cornerstone of digital media, from entertainment and education to professional communication.
no code implementations • 12 Jun 2024 • Bing Li, Cheng Zheng, Wenxuan Zhu, Jinjie Mai, Biao Zhang, Peter Wonka, Bernard Ghanem
To this end, we propose a novel diffusion-based pipeline that generates high-quality multi-view videos centered around a dynamic 3D object from text.
2 code implementations • 10 Jun 2024 • Zhenyu Li, Shariq Farooq Bhat, Peter Wonka
This paper introduces PatchRefiner, an advanced framework for metric single image depth estimation aimed at high-resolution real-domain inputs.
no code implementations • 1 Jun 2024 • Hanxiao Wang, Mingyang Zhao, Weize Quan, Zhen Chen, Dong-Ming Yan, Peter Wonka
To address this issue, we propose E3-Net to achieve equivariance for normal estimation.
1 code implementation • 27 May 2024 • Qian Wang, Abdelrahman Eldesokey, Mohit Mendiratta, Fangneng Zhan, Adam Kortylewski, Christian Theobalt, Peter Wonka
We introduce the first zero-shot approach for Video Semantic Segmentation (VSS) based on pre-trained diffusion models.
no code implementations • 24 May 2024 • Bingchen Yang, Haiyong Jiang, Hao Pan, Peter Wonka, Jun Xiao, Guosheng Lin
At each step, we provide two forms of geometric guidance.
no code implementations • 19 Mar 2024 • Yazeed Alharbi, Peter Wonka
We present a novel, training-free approach for textual editing of real images using diffusion models.
no code implementations • 8 Feb 2024 • Wamiq Reyaz Para, Abdelrahman Eldesokey, Zhenyu Li, Pradyumna Reddy, Jiankang Deng, Peter Wonka
To the best of our knowledge, our approach is the first to introduce multi-modal conditioning to 3D avatar generation and editing.
no code implementations • 7 Jan 2024 • Weize Quan, Jiaxi Chen, Yanli Liu, Dong-Ming Yan, Peter Wonka
The goal of this paper is to comprehensively review the deep learning-based methods for image and video inpainting.
no code implementations • CVPR 2024 • Tom Kelly, John Femiani, Peter Wonka
We evaluate a procedural model by training semantic segmentation networks on both synthetic and real images and then comparing their performances on a shared test set of real images.
1 code implementation • 13 Dec 2023 • Mykola Lavreniuk, Shariq Farooq Bhat, Matthias Müller, Peter Wonka
Second, we propose a novel image-text alignment module for improved feature extraction of the Stable Diffusion backbone.
Ranked #1 on Depth Estimation on NYU-Depth V2
1 code implementation • 12 Dec 2023 • Abdelrahman Eldesokey, Peter Wonka
We strive to bridge this gap, and we introduce LatentMan, which leverages existing text-based motion diffusion models to generate diverse continuous motions to guide the T2I model.
1 code implementation • 11 Dec 2023 • Jian Shi, Peter Wonka
To the best of our knowledge, \textit{VoxelKP} is the first single-staged, fully sparse network that is specifically designed for addressing the challenging task of 3D keypoint estimation from LiDAR data, achieving state-of-the-art performances.
Ranked #1 on 3D Human Pose Estimation on Waymo Open Dataset
no code implementations • 7 Dec 2023 • Savva Ignatyev, Daniil Selikhanovych, Oleg Voynov, Yiqun Wang, Peter Wonka, Stamatios Lefkimmiatis, Evgeny Burnaev
We present a novel method for 3D surface reconstruction from multiple images where only a part of the object of interest is captured.
no code implementations • 5 Dec 2023 • Shariq Farooq Bhat, Niloy J. Mitra, Peter Wonka
We present LooseControl to allow generalized depth conditioning for diffusion-based image generation.
1 code implementation • CVPR 2024 • Zhenyu Li, Shariq Farooq Bhat, Peter Wonka
Single image depth estimation is a foundational task in computer vision and generative modeling.
1 code implementation • CVPR 2024 • Sherwin Bahmani, Ivan Skorokhodov, Victor Rong, Gordon Wetzstein, Leonidas Guibas, Peter Wonka, Sergey Tulyakov, Jeong Joon Park, Andrea Tagliasacchi, David B. Lindell
Recent breakthroughs in text-to-4D generation rely on pre-trained text-to-image and text-to-video models to generate dynamic 3D scenes.
1 code implementation • CVPR 2024 • Thomas Wimmer, Peter Wonka, Maks Ovsjanikov
In this work, we propose to explore foundation models for the task of keypoint detection on 3D shapes.
no code implementations • CVPR 2024 • Biao Zhang, Peter Wonka
We propose a new class of generative diffusion models, called functional diffusion.
1 code implementation • 27 Oct 2023 • Habib Slim, Xiang Li, Yuchen Li, Mahmoud Ahmed, Mohamed Ayman, Ujjwal Upadhyay, Ahmed Abdelreheem, Arpit Prajapati, Suhail Pothigara, Peter Wonka, Mohamed Elhoseiny
In this work, we present 3DCoMPaT$^{++}$, a multimodal 2D/3D dataset with 160 million rendered views of more than 10 million stylized 3D shapes carefully annotated at the part-instance level, alongside matching RGB point clouds, 3D textured meshes, depth maps, and segmentation masks.
1 code implementation • 16 Oct 2023 • Hanan Gani, Shariq Farooq Bhat, Muzammal Naseer, Salman Khan, Peter Wonka
Diffusion-based generative models have significantly advanced text-to-image generation but encounter challenges when processing lengthy and intricate text prompts describing complex scenes with multiple objects.
no code implementations • 11 Oct 2023 • Ryan Po, Wang Yifan, Vladislav Golyanik, Kfir Aberman, Jonathan T. Barron, Amit H. Bermano, Eric Ryan Chan, Tali Dekel, Aleksander Holynski, Angjoo Kanazawa, C. Karen Liu, Lingjie Liu, Ben Mildenhall, Matthias Nießner, Björn Ommer, Christian Theobalt, Peter Wonka, Gordon Wetzstein
The field of visual computing is rapidly advancing due to the emergence of generative artificial intelligence (AI), which unlocks unprecedented capabilities for the generation, editing, and reconstruction of images, videos, and 3D scenes.
1 code implementation • 9 Oct 2023 • Tom Kelly, John Femiani, Peter Wonka
We evaluate a procedural model by training semantic segmentation networks on both synthetic and real images and then comparing their performances on a shared test set of real images.
1 code implementation • 30 Jun 2023 • Guocheng Qian, Jinjie Mai, Abdullah Hamdi, Jian Ren, Aliaksandr Siarohin, Bing Li, Hsin-Ying Lee, Ivan Skorokhodov, Peter Wonka, Sergey Tulyakov, Bernard Ghanem
We present Magic123, a two-stage coarse-to-fine approach for high-quality, textured 3D meshes generation from a single unposed image in the wild using both2D and 3D priors.
no code implementations • 5 Jun 2023 • Ahmed Abdelreheem, Abdelrahman Eldesokey, Maks Ovsjanikov, Peter Wonka
Instead, we propose to exploit the in-context learning capabilities of ChatGPT to generate two different sets of semantic regions for each shape and a semantic mapping between them.
1 code implementation • 29 May 2023 • Qian Wang, Biao Zhang, Michael Birsak, Peter Wonka
In this work, we propose a framework termed InstructEdit that can do fine-grained editing based on user instructions.
no code implementations • 29 May 2023 • Yue Fan, Ivan Skorokhodov, Oleg Voynov, Savva Ignatyev, Evgeny Burnaev, Peter Wonka, Yiqun Wang
We develop a method that recovers the surface, materials, and illumination of a scene from its posed multi-view images.
1 code implementation • CVPR 2023 • Yiqun Wang, Ivan Skorokhodov, Peter Wonka
The first component is to borrow the tri-plane representation from EG3D and represent signed distance fields as a mixture of tri-planes and MLPs instead of representing it with MLPs only.
no code implementations • ICCV 2023 • Ahmed Abdelreheem, Ivan Skorokhodov, Maks Ovsjanikov, Peter Wonka
We explore the task of zero-shot semantic segmentation of 3D shapes by using large-scale off-the-shelf 2D image recognition models.
2 code implementations • 29 Mar 2023 • Qian Wang, Biao Zhang, Michael Birsak, Peter Wonka
Image generation using diffusion can be controlled in multiple ways.
1 code implementation • CVPR 2023 • Anna Frühstück, Nikolaos Sarafianos, Yuanlu Xu, Peter Wonka, Tony Tung
Our experiments demonstrate that VIVE3D generates high-fidelity face edits at consistent quality from a range of camera viewpoints which are composited with the original video in a temporally and spatially consistent manner.
no code implementations • 26 Mar 2023 • Qian Wang, Yiqun Wang, Michael Birsak, Peter Wonka
3D-aware image synthesis has attracted increasing interest as it models the 3D nature of our real world.
no code implementations • 2 Mar 2023 • Ivan Skorokhodov, Aliaksandr Siarohin, Yinghao Xu, Jian Ren, Hsin-Ying Lee, Peter Wonka, Sergey Tulyakov
Existing 3D-from-2D generators are typically designed for well-curated single-category datasets, where all the objects have (approximately) the same scale, 3D location, and orientation, and the camera always points to the center of the scene.
1 code implementation • 28 Feb 2023 • Jian Shi, Pengyi Zhang, Ni Zhang, Hakim Ghazzai, Peter Wonka
DIA is a fine-grained anomaly detection framework for medical images.
5 code implementations • 23 Feb 2023 • Shariq Farooq Bhat, Reiner Birkl, Diana Wofk, Peter Wonka, Matthias Müller
Finally, ZoeD-M12-NK is the first model that can jointly train on multiple datasets (NYU Depth v2 and KITTI) without a significant drop in performance and achieve unprecedented zero-shot generalization performance to eight unseen datasets from both indoor and outdoor domains.
Ranked #23 on Monocular Depth Estimation on NYU-Depth V2 (using extra training data)
1 code implementation • 26 Jan 2023 • Biao Zhang, Jiapeng Tang, Matthias Niessner, Peter Wonka
We introduce 3DShape2VecSet, a novel shape representation for neural fields designed for generative diffusion models.
no code implementations • CVPR 2023 • Rameen Abdal, Hsin-Ying Lee, Peihao Zhu, Menglei Chai, Aliaksandr Siarohin, Peter Wonka, Sergey Tulyakov
Finally, we propose a novel inversion method for 3D-GANs linking the latent spaces of the source and the target domains.
no code implementations • 12 Dec 2022 • Ahmed Abdelreheem, Kyle Olszewski, Hsin-Ying Lee, Peter Wonka, Panos Achlioptas
The two popular datasets ScanRefer [16] and ReferIt3D [3] connect natural language to real-world 3D data.
no code implementations • 20 Oct 2022 • Jheng-Wei Su, Chi-Han Peng, Peter Wonka, Hung-Kuo Chu
The major improvement over PSMNet comes from a novel Geometry-aware Panorama Registration Network or GPR-Net that effectively tackles the wide baseline registration problem by exploiting the layout geometry and computing fine-grained correspondences on the layout boundaries, instead of the global pixel-space.
1 code implementation • 1 Sep 2022 • Michael Birsak, Tom Kelly, Wamiq Para, Peter Wonka
We present a novel generative method for the creation of city-scale road layouts.
1 code implementation • 25 Aug 2022 • Yicheng Luo, Jing Ren, Xuefei Zhe, Di Kang, Yajing Xu, Peter Wonka, Linchao Bao
The network takes a line cloud as input , i. e., a nonstructural and unordered set of 3D line segments extracted from multi-view images, and outputs a 3D wireframe of the underlying building, which consists of a sparse set of 3D junctions connected by line segments.
no code implementations • 25 Aug 2022 • Ildar Gilmutdinov, Ingrid Schloegel, Alois Hinterleitner, Peter Wonka, Michael Wimmer
Assessing the structure of a building with non-invasive methods is an important problem.
1 code implementation • 21 Jun 2022 • Ivan Skorokhodov, Sergey Tulyakov, Yiqun Wang, Peter Wonka
In this work, we show that it is possible to obtain a high-resolution 3D generator with SotA image quality by following a completely different route of simply training the model patch-wise.
1 code implementation • 15 Jun 2022 • Yiqun Wang, Ivan Skorokhodov, Peter Wonka
We develop HF-NeuS, a novel method to improve the quality of surface reconstruction in neural rendering.
no code implementations • 15 Jun 2022 • Abdalla G. M. Ahmed, Jing Ren, Peter Wonka
Among the various approaches for producing point distributions with blue noise spectrum, we argue for an optimization framework using Gaussian kernels.
no code implementations • 1 Jun 2022 • Azimkhon Ostonov, Peter Wonka, Dominik L. Michels
We present RLSS: a reinforcement learning algorithm for sequential scene generation.
no code implementations • 29 May 2022 • Wamiq Reyaz Para, Paul Guerrero, Niloy Mitra, Peter Wonka
Scalable generation of furniture layouts is essential for many applications in virtual reality, augmented reality, game development and synthetic data generation.
no code implementations • 27 May 2022 • Rameen Abdal, Peihao Zhu, Niloy J. Mitra, Peter Wonka
Image editing using a pretrained StyleGAN generator has emerged as a powerful paradigm for facial editing, providing disentangled controls over age, expression, illumination, etc.
1 code implementation • 27 May 2022 • Biao Zhang, Matthias Nießner, Peter Wonka
All probabilistic experiments confirm that we are able to generate detailed and high quality shapes to yield the new state of the art in generative 3D shape modeling.
no code implementations • 8 May 2022 • Ngan Nguyen, Feng Liang, Dominik Engel, Ciril Bohak, Peter Wonka, Timo Ropinski, Ivan Viola
We propose a new microscopy simulation system that can depict atomistic models in a micrograph visual style, similar to results of physical electron microscopy imaging.
1 code implementation • 28 Mar 2022 • Shariq Farooq Bhat, Ibraheem Alhashim, Peter Wonka
We build on AdaBins which estimates a global distribution of depth values for the input image and evolve the architecture in two ways.
Ranked #43 on Monocular Depth Estimation on NYU-Depth V2
2 code implementations • CVPR 2022 • Anna Frühstück, Krishna Kumar Singh, Eli Shechtman, Niloy J. Mitra, Peter Wonka, Jingwan Lu
Instead of modeling this complex domain with a single GAN, we propose a novel method to combine multiple pretrained GANs, where one GAN generates a global canvas (e. g., human body) and a set of specialized GANs, or insets, focus on different parts (e. g., faces, shoes) that can be seamlessly inserted onto the global canvas.
1 code implementation • 31 Jan 2022 • Motasem Alfarra, Juan C. Pérez, Anna Frühstück, Philip H. S. Torr, Peter Wonka, Bernard Ghanem
Finally, we show that the FID can be robustified by simply replacing the standard Inception with a robust Inception.
no code implementations • 9 Dec 2021 • Rameen Abdal, Peihao Zhu, John Femiani, Niloy J. Mitra, Peter Wonka
The success of StyleGAN has enabled unprecedented semantic editing capabilities, on both synthesized and real images.
2 code implementations • ICLR 2022 • Peihao Zhu, Rameen Abdal, John Femiani, Peter Wonka
The input to our method is trained GAN that can produce images in domain A and a single reference image I_B from domain B.
no code implementations • ICCV 2021 • Dong Lao, Peihao Zhu, Peter Wonka, Ganesh Sundaramoorthi
We consider the problem of filling in missing spatio-temporal regions of a video.
no code implementations • CVPR 2021 • Gautam Pai, Jing Ren, Simone Melzi, Peter Wonka, Maks Ovsjanikov
In this paper, we provide a theoretical foundation for pointwise map recovery from functional maps and highlight its relation to a range of shape correspondence methods based on spectral alignment.
no code implementations • NeurIPS 2021 • Wamiq Reyaz Para, Shariq Farooq Bhat, Paul Guerrero, Tom Kelly, Niloy Mitra, Leonidas Guibas, Peter Wonka
Sketches can be represented as graphs, with the primitives as nodes and the constraints as edges.
no code implementations • 4 Jun 2021 • Hiroyasu Akada, Shariq Farooq Bhat, Ibraheem Alhashim, Peter Wonka
Specifically, we extend self-supervised learning from traditional representation learning, which works on images from a single domain, to domain invariant representation learning, which works on images from two different domains by utilizing an image-to-image translation network.
1 code implementation • 2 Jun 2021 • Peihao Zhu, Rameen Abdal, John Femiani, Peter Wonka
Seamlessly blending features from multiple images is extremely challenging because of complex relationships in lighting, geometry, and partial occlusion which cause coupling between different parts of the image.
no code implementations • 4 Apr 2021 • Ngan Nguyen, Ciril Bohak, Dominik Engel, Peter Mindek, Ondřej Strnad, Peter Wonka, Sai Li, Timo Ropinski, Ivan Viola
Our technique shows the high impact in target sciences for visual data analysis of very noisy volumes that cannot be visualized with existing techniques.
1 code implementation • ICCV 2021 • Rameen Abdal, Peihao Zhu, Niloy Mitra, Peter Wonka
We propose an unsupervised segmentation framework for StyleGAN generated objects.
1 code implementation • ICCV 2021 • Guangming Zang, Ramzi Idoughi, Rui Li, Peter Wonka, Wolfgang Heidrich
After getting estimated through the sinogram prediction module, the density field is consistently refined in the second module using local and non-local geometrical priors.
Ranked #6 on Novel View Synthesis on X3D
Computed Tomography (CT) Low-Dose X-Ray Ct Reconstruction +3
3 code implementations • 13 Dec 2020 • Peihao Zhu, Rameen Abdal, Yipeng Qin, John Femiani, Peter Wonka
First, we introduce a new normalized space to analyze the diversity and the quality of the reconstructed latent codes.
11 code implementations • CVPR 2021 • Shariq Farooq Bhat, Ibraheem Alhashim, Peter Wonka
We address the problem of estimating a high quality dense depth map from a single RGB input image.
Ranked #7 on Depth Estimation on NYU-Depth V2
no code implementations • ICCV 2021 • Wamiq Para, Paul Guerrero, Tom Kelly, Leonidas Guibas, Peter Wonka
We generate layouts in three steps.
no code implementations • ICLR 2022 • Biao Zhang, Peter Wonka
We combine training data generating networks with bi-level optimization algorithms to obtain a complete framework for which all components can be jointly trained.
no code implementations • 25 Aug 2020 • Dong Lao, Peihao Zhu, Peter Wonka, Ganesh Sundaramoorthi
We introduce optimization methods for convolutional neural networks that can be used to improve existing gradient-based optimization in terms of generalization error.
3 code implementations • 6 Aug 2020 • Rameen Abdal, Peihao Zhu, Niloy Mitra, Peter Wonka
We evaluate our method using the face and the car latent space of StyleGAN, and demonstrate fine-grained disentangled edits along various attributes on both real photographs and StyleGAN generated images.
1 code implementation • CVPR 2020 • Yazeed Alharbi, Peter Wonka
We show that disentanglement in the first layer of the generator network leads to disentanglement in the generated image.
no code implementations • 28 Jan 2020 • Yiqun Wang, Jing Ren, Dong-Ming Yan, Jianwei Guo, Xiaopeng Zhang, Peter Wonka
Second, we propose a new multiscale graph convolutional network (MGCN) to transform a non-learned feature to a more discriminative descriptor.
no code implementations • CVPR 2021 • Biao Zhang, Peter Wonka
In this paper we propose a new framework for point cloud instance segmentation.
Ranked #1 on Instance Segmentation on PartNet
1 code implementation • CVPR 2020 • Peihao Zhu, Rameen Abdal, Yipeng Qin, Peter Wonka
Using SEAN normalization, we can build a network architecture that can control the style of each semantic region individually, e. g., we can specify one style reference image per region.
3 code implementations • CVPR 2020 • Rameen Abdal, Yipeng Qin, Peter Wonka
We propose Image2StyleGAN++, a flexible image editing framework with many applications.
1 code implementation • CVPR 2020 • Kaichun Mo, Paul Guerrero, Li Yi, Hao Su, Peter Wonka, Niloy Mitra, Leonidas J. Guibas
Learning to encode differences in the geometry and (topological) structure of the shapes of ordinary objects is key to generating semantically plausible variations of a given shape, transferring edits from one shape to another, and many other applications in 3D content creation.
3 code implementations • 9 Oct 2019 • Chuhang Zou, Jheng-Wei Su, Chi-Han Peng, Alex Colburn, Qi Shan, Peter Wonka, Hung-Kuo Chu, Derek Hoiem
Recent approaches for predicting layouts from 360 panoramas produce excellent results.
no code implementations • 12 Sep 2019 • Károly Zsolnai-Fehér, Peter Wonka, Michael Wimmer
In the proposed workflow, the user starts with an input image and applies a few intuitive transforms (e. g., colorization, image inpainting) within a 2D image editor of their choice, and in the next step, our technique produces a photorealistic result that approximates this target image.
2 code implementations • 1 Aug 2019 • Kaichun Mo, Paul Guerrero, Li Yi, Hao Su, Peter Wonka, Niloy Mitra, Leonidas J. Guibas
We introduce StructureNet, a hierarchical graph network which (i) can directly encode shapes represented as such n-ary graphs; (ii) can be robustly trained on large and complex shape families; and (iii) can be used to generate a great diversity of realistic structured shape geometries.
1 code implementation • 29 Apr 2019 • Anna Frühstück, Ibraheem Alhashim, Peter Wonka
We tackle the problem of texture synthesis in the setting where many input images are given and a large-scale output is required.
2 code implementations • 16 Apr 2019 • Simone Melzi, Jing Ren, Emanuele Rodolà, Abhishek Sharma, Peter Wonka, Maks Ovsjanikov
Our main observation is that high quality maps can be obtained even if the input correspondences are noisy or are encoded by a small number of coefficients in a spectral basis.
Graphics
5 code implementations • ICCV 2019 • Rameen Abdal, Yipeng Qin, Peter Wonka
We propose an efficient algorithm to embed a given image into the latent space of StyleGAN.
45 code implementations • 31 Dec 2018 • Ibraheem Alhashim, Peter Wonka
Accurate depth estimation from images is a fundamental task in many applications including scene understanding and reconstruction.
Ranked #49 on Monocular Depth Estimation on KITTI Eigen split
no code implementations • CVPR 2019 • Yazeed Alharbi, Neil Smith, Peter Wonka
In multimodal unsupervised image-to-image translation tasks, the goal is to translate an image from the source domain to many images in the target domain.
1 code implementation • CVPR 2019 • Shang-Ta Yang, Fu-En Wang, Chi-Han Peng, Peter Wonka, Min Sun, Hung-Kuo Chu
We present a deep learning framework, called DuLa-Net, to predict Manhattan-world 3D room layouts from a single RGB panorama.
no code implementations • ECCV 2020 • Yipeng Qin, Niloy Mitra, Peter Wonka
In this work, we uncover an even more important effect of Lipschitz regularization by examining its impact on the loss function: It degenerates GAN loss functions to almost linear ones by restricting their domain and interval of attainable gradient values.
1 code implementation • SIGGRAPH Asia 2018 • Tom Kelly, Paul Guerrero, Anthony Steed, Peter Wonka, Niloy J. Mitra
The various GANs are synchronized to produce consistent style distributions over buildings and neighborhoods.
no code implementations • ECCV 2018 • Guangming Zang, Mohamed Aly, Ramzi Idoughi, Peter Wonka, Wolfgang Heidrich
As a second, smaller contribution, we also show that when using such a proximal reconstruction framework, it is beneficial to employ the Simultaneous Algebraic Reconstruction Technique (SART) instead of the commonly used Conjugate Gradient (CG) method in the solution of the data term proximal operator.
2 code implementations • 19 Jun 2018 • Tom Kelly, Paul Guerrero, Anthony Steed, Peter Wonka, Niloy J. Mitra
The various GANs are synchronized to produce consistent style distributions over buildings and neighborhoods.
Graphics
no code implementations • 9 May 2018 • John Femiani, Wamiq Reyaz Para, Niloy Mitra, Peter Wonka
Specifically, we propose a MULTIFACSEGNET architecture to assign multiple labels to each pixel, a SEPARABLE architecture as a low-rank formulation that encourages extraction of rectangular elements, and a COMPATIBILITY network that simultaneously seeks segmentation across facade element types allowing the network to 'see' intermediate output probabilities of the various facade element classes.
no code implementations • 23 Apr 2018 • Károly Zsolnai-Fehér, Peter Wonka, Michael Wimmer
Workflow timings against Disney's "principled" shader reveal that our system scales well with the number of sought materials, thus empowering even novice users to generate hundreds of high-quality material models without any expertise in material modeling.
no code implementations • 8 Apr 2018 • Lama Affara, Bernard Ghanem, Peter Wonka
Convolutional Sparse Coding (CSC) is a well-established image representation model especially suited for image restoration tasks.
no code implementations • 12 Feb 2018 • Silvio Giancola, Jens Schneider, Peter Wonka, Bernard S. Ghanem
We also present a technique to filter the pairs of 3D matched points based on the distribution of their distances.
no code implementations • ICCV 2017 • Liangliang Nan, Peter Wonka
We show that reconstruction from point clouds can be cast as a binary labeling problem.
no code implementations • 27 Sep 2017 • Lama Affara, Bernard Ghanem, Peter Wonka
Convolutional sparse coding (CSC) is an important building block of many computer vision applications ranging from image and video compression to deep learning.
1 code implementation • 11 Jun 2016 • Mohamed Aly, Guangming Zang, Wolfgang Heidrich, Peter Wonka
We present TRex, a flexible and robust Tomographic Reconstruction framework using proximal algorithms.
no code implementations • ICCV 2015 • Mohammed Hachama, Bernard Ghanem, Peter Wonka
In this paper, we address the problem of computing an intrinsic decomposition of the colors of a surface into an albedo and a shading term.
no code implementations • 5 Oct 2015 • Paul Guerrero, Niloy J. Mitra, Peter Wonka
As humans, we regularly interpret images based on the relations between image regions.
no code implementations • 25 Oct 2013 • Jie Wang, Peter Wonka, Jieping Ye
Some appealing features of our screening method are: (1) DVI is safe in the sense that the vectors discarded by DVI are guaranteed to be non-support vectors; (2) the data set needs to be scanned only once to run the screening, whose computational cost is negligible compared to that of solving the SVM problem; (3) DVI is independent of the solvers and can be integrated with any existing efficient solvers.
no code implementations • NeurIPS 2014 • Jie Wang, Jiayu Zhou, Jun Liu, Peter Wonka, Jieping Ye
The l1-regularized logistic regression (or sparse logistic regression) is a widely used method for simultaneous classification and feature selection.
no code implementations • NeurIPS 2013 • Jie Wang, Peter Wonka, Jieping Ye
To improve the efficiency of solving large-scale Lasso problems, El Ghaoui and his colleagues have proposed the SAFE rules which are able to quickly identify the inactive predictors, i. e., predictors that have $0$ components in the solution vector.
no code implementations • 10 Sep 2012 • Sen Yang, Zhaosong Lu, Xiaotong Shen, Peter Wonka, Jieping Ye
We expect the two brain networks for NC and MCI to share common structures but not to be identical to each other; similarly for the two brain networks for MCI and AD.
no code implementations • NeurIPS 2010 • Ji Liu, Peter Wonka, Jieping Ye
We show that if $X$ obeys a certain condition, then with a large probability the difference between the solution $\hat\beta$ estimated by the proposed method and the true solution $\beta^*$ measured in terms of the $l_p$ norm ($p\geq 1$) is bounded as \begin{equation*} \|\hat\beta-\beta^*\|_p\leq \left(C(s-N)^{1/p}\sqrt{\log m}+\Delta\right)\sigma, \end{equation*} $C$ is a constant, $s$ is the number of nonzero entries in $\beta^*$, $\Delta$ is independent of $m$ and is much smaller than the first term, and $N$ is the number of entries of $\beta^*$ larger than a certain value in the order of $\mathcal{O}(\sigma\sqrt{\log m})$.