no code implementations • 3 Mar 2025 • Juil Koo, Paul Guerrero, Chun-Hao Paul Huang, Duygu Ceylan, Minhyuk Sung
To the best of our knowledge, this is the first generative approach to edit object compositions in videos.
no code implementations • 2 Jan 2025 • Chun-Hao Paul Huang, Jae Shin Yoon, Hyeonho Jeong, Niloy Mitra, Duygu Ceylan
Inspired by the emergent 3D capabilities in image generators, we explore whether video generators similarly exhibit 3D awareness.
no code implementations • 23 Dec 2024 • Fa-Ting Hong, Zhan Xu, Haiyang Liu, Qinjie Lin, Luchuan Song, Zhixin Shu, Yang Zhou, Duygu Ceylan, Dan Xu
Diffusion-based human animation aims to animate a human character based on a source human image as well as driving signals such as a sequence of poses.
no code implementations • 21 Dec 2024 • Souhaib Attaiki, Paul Guerrero, Duygu Ceylan, Niloy J. Mitra, Maks Ovsjanikov
We observe that GAN- and diffusion-based generators have complementary qualities: GANs can be trained efficiently with 2D supervision to produce high-quality 3D objects but are hard to condition on text.
no code implementations • 8 Dec 2024 • Hyeonho Jeong, Chun-Hao Paul Huang, Jong Chul Ye, Niloy Mitra, Duygu Ceylan
While recent foundational video generators produce visually rich output, they still struggle with appearance drift, where objects gradually degrade or change inconsistently across frames, breaking visual coherence.
no code implementations • 22 Nov 2024 • Swasti Shreya Mishra, Kuldeep Kulkarni, Duygu Ceylan, Balaji Vasan Srinivasan
The input to our model is a text prompt depicting the type of clothing and the texture of clothing like leopard, striped, or plain, and a sequence of normal maps that capture the underlying animation that we desire in the output.
1 code implementation • 19 Nov 2024 • Abdul Basit Anees, Ahmet Canberk Baykal, Muhammed Burak Kizil, Duygu Ceylan, Erkut Erdem, Aykut Erdem
Towards this end, in this study, we present a novel framework that significantly extends the capabilities of a pre-trained StyleGAN by integrating CLIP space via hypernetworks.
no code implementations • 14 Oct 2024 • Soon Yau Cheong, Duygu Ceylan, Armin Mustafa, Andrew Gilbert, Chun-Hao Paul Huang
While U-Net-based models have shown promising results for camera control, transformer-based diffusion models (DiT)-the preferred architecture for large-scale video generation - suffer from severe degradation in camera motion accuracy.
no code implementations • 31 Jul 2024 • Doğa Yılmaz, Towaki Takikawa, Duygu Ceylan, Kaan Akşit
Uniquely, a single inference step of our model supports different permutations of these perceptual tasks at different prompted rates (i. e., mildly, lightly), eliminating the need for daisy-chaining multiple models to get the desired perceptual effect.
no code implementations • 2 Jun 2024 • Yuan Shen, Duygu Ceylan, Paul Guerrero, Zexiang Xu, Niloy J. Mitra, Shenlong Wang, Anna Frühstück
We demonstrate that it is possible to directly repurpose existing (pretrained) video models for 3D super-resolution and thus sidestep the problem of the shortage of large repositories of high-quality 3D training models.
1 code implementation • 13 May 2024 • Peizhuo Li, Tuanfeng Y. Wang, Timur Levent Kesdogan, Duygu Ceylan, Olga Sorkine-Hornung
Data driven and learning based solutions for modeling dynamic garments have significantly advanced, especially in the context of digital humans.
no code implementations • 1 May 2024 • Burak Can Biner, Farrin Marouf Sofian, Umur Berkay Karakaş, Duygu Ceylan, Erkut Erdem, Aykut Erdem
In addition to audio conditioned image generation, our method can also be utilized in conjuction with diffusion based editing methods to enable audio conditioned image editing.
no code implementations • 3 Apr 2024 • Duygu Ceylan, Valentin Deschaintre, Thibault Groueix, Rosalie Martin, Chun-Hao Huang, Romain Rouffet, Vladimir Kim, Gaëtan Lassagne
We present MatAtlas, a method for consistent text-guided 3D model texturing.
no code implementations • CVPR 2024 • Shengqu Cai, Duygu Ceylan, Matheus Gadelha, Chun-Hao Paul Huang, Tuanfeng Yang Wang, Gordon Wetzstein
Traditional 3D content creation tools empower users to bring their imagination to life by giving them direct control over a scene's geometry, appearance, motion, and camera path.
no code implementations • 4 Sep 2023 • Sanjeev Muralikrishnan, Chun-Hao Paul Huang, Duygu Ceylan, Niloy J. Mitra
Morphable models are fundamental to numerous human-centered processes as they offer a simple yet expressive shape space.
no code implementations • 22 Aug 2023 • Omid Taheri, Yi Zhou, Dimitrios Tzionas, Yang Zhou, Duygu Ceylan, Soren Pirk, Michael J. Black
In contrast, we introduce GRIP, a learning-based method that takes, as input, the 3D motion of the body and the object, and synthesizes realistic motion for both hands before, during, and after object interaction.
no code implementations • 17 Jul 2023 • Ahmet Canberk Baykal, Abdul Basit Anees, Duygu Ceylan, Erkut Erdem, Aykut Erdem, Deniz Yuret
Existing approaches for editing images using language either resort to instance-level latent code optimization or map predefined text prompts to some editing directions in the latent space.
no code implementations • ICCV 2023 • Moayed Haji Ali, Andrew Bond, Tolga Birdal, Duygu Ceylan, Levent Karacan, Erkut Erdem, Aykut Erdem
However, the applicability of such advancements to the video domain has been hindered by the difficulty of representing and controlling videos in the latent space of GANs.
no code implementations • 10 Apr 2023 • Youngjoong Kwon, Dahun Kim, Duygu Ceylan, Henry Fuchs
We present a method that enables synthesizing novel views and novel poses of arbitrary human performers from sparse multi-view images.
1 code implementation • ICCV 2023 • Duygu Ceylan, Chun-Hao Paul Huang, Niloy J. Mitra
Our method works in two simple steps: first, we use a pre-trained structure-guided (e. g., depth) image diffusion model to perform text-guided edits on an anchor frame; then, in the key step, we progressively propagate the changes to the future frames via self-attention feature injection to adapt the core denoising step of the diffusion model.
no code implementations • CVPR 2023 • Hugo Bertiche, Niloy J. Mitra, Kuldeep Kulkarni, Chun-Hao Paul Huang, Tuanfeng Y. Wang, Meysam Madadi, Sergio Escalera, Duygu Ceylan
We investigate the problem in the context of dressed humans under the wind.
no code implementations • CVPR 2023 • Yasamin Jafarian, Tuanfeng Y. Wang, Duygu Ceylan, Jimei Yang, Nathan Carr, Yi Zhou, Hyun Soo Park
To edit human videos in a physically plausible way, a texture map must take into account not only the garment transformation induced by the body movements and clothes fitting, but also its 3D fine-grained surface geometry.
no code implementations • ICCV 2023 • Rishabh Jain, Mayur Hemani, Duygu Ceylan, Krishna Kumar Singh, Jingwan Lu, Mausoom Sarkar, Balaji Krishnamurthy
Numerous pose-guided human editing methods have been explored by the vision community due to their extensive practical applications.
no code implementations • CVPR 2023 • Rishabh Jain, Krishna Kumar Singh, Mayur Hemani, Jingwan Lu, Mausoom Sarkar, Duygu Ceylan, Balaji Krishnamurthy
The task of human reposing involves generating a realistic image of a person standing in an arbitrary conceivable pose.
1 code implementation • 23 Sep 2022 • Meng Zhang, Duygu Ceylan, Niloy J. Mitra
Technically, we model garment dynamics, driven using the input character motion, by predicting per-frame local displacements in a canonical state of the garment that is enriched with frame-dependent skinning weights to bring the garment to the global space.
1 code implementation • 23 Aug 2022 • Chun-Han Yao, Jimei Yang, Duygu Ceylan, Yi Zhou, Yang Zhou, Ming-Hsuan Yang
An alternative approach is to estimate dense vertices of a predefined template body in the image space.
no code implementations • 28 Jul 2022 • Qingyang Tan, Yi Zhou, Tuanfeng Wang, Duygu Ceylan, Xin Sun, Dinesh Manocha
Despite recent success, deep learning-based methods for predicting 3D garment deformation under body motion suffer from interpenetration problems between the garment and the body.
no code implementations • 14 May 2022 • Yunseok Jang, Ruben Villegas, Jimei Yang, Duygu Ceylan, Xin Sun, Honglak Lee
We test the effectiveness of our representation on the human image harmonization task by predicting shading that is coherent with a given background image.
no code implementations • CVPR 2022 • Jae Shin Yoon, Duygu Ceylan, Tuanfeng Y. Wang, Jingwan Lu, Jimei Yang, Zhixin Shu, Hyun Soo Park
Appearance of dressed humans undergoes a complex geometric transformation induced not only by the static pose but also by its dynamics, i. e., there exists a number of cloth geometric configurations given a pose depending on the way it has moved.
no code implementations • 10 Nov 2021 • Tuanfeng Y. Wang, Duygu Ceylan, Krishna Kumar Singh, Niloy J. Mitra
Synthesizing dynamic appearances of humans in motion plays a central role in applications such as AR/VR and video editing.
1 code implementation • NeurIPS 2021 • Youngjoong Kwon, Dahun Kim, Duygu Ceylan, Henry Fuchs
To tackle this, we propose Neural Human Performer, a novel approach that learns generalizable neural radiance fields based on a parametric human body model for robust performance capture.
Ranked #3 on
Generalizable Novel View Synthesis
on ZJU-MoCap
no code implementations • ICCV 2021 • Ruben Villegas, Duygu Ceylan, Aaron Hertzmann, Jimei Yang, Jun Saito
Self-contacts, such as when hands touch each other or the torso or the head, are important attributes of human body language and dynamics, yet existing methods do not model or preserve these contacts.
1 code implementation • ICCV 2021 • Eric-Tuan Lê, Minhyuk Sung, Duygu Ceylan, Radomir Mech, Tamy Boubekeur, Niloy J. Mitra
We present Cascaded Primitive Fitting Networks (CPFN) that relies on an adaptive patch sampling network to assemble detection results of global and local primitive detection networks.
no code implementations • ICCV 2021 • Mohamed Hassan, Duygu Ceylan, Ruben Villegas, Jun Saito, Jimei Yang, Yi Zhou, Michael Black
A long-standing goal in computer vision is to capture, model, and realistically synthesize human behavior.
no code implementations • 7 Jun 2021 • Jiaman Li, Ruben Villegas, Duygu Ceylan, Jimei Yang, Zhengfei Kuang, Hao Li, Yajie Zhao
We demonstrate the effectiveness of our hierarchical motion variational autoencoder in a variety of tasks including video-based human pose estimation, motion completion from partial observations, and motion synthesis from sparse key-frames.
Ranked #4 on
Motion Synthesis
on LaFAN1
no code implementations • CVPR 2021 • Mianlun Zheng, Yi Zhou, Duygu Ceylan, Jernej Barbič
Being a local method, our network is independent of the mesh topology and generalizes to arbitrarily shaped 3D character meshes at test time.
no code implementations • 23 Feb 2021 • Meng Zhang, Duygu Ceylan, Tuanfeng Wang, Niloy J. Mitra
A vital task of the wider digital human effort is the creation of realistic garments on digital avatars, both in the form of characteristic fold patterns and wrinkles in static frames as well as richness of garment dynamics under avatars' motion.
1 code implementation • CVPR 2020 • Kyle Olszewski, Duygu Ceylan, Jun Xing, Jose Echevarria, Zhili Chen, Weikai Chen, Hao Li
We present an interactive approach to synthesizing realistic variations in facial hair in images, ranging from subtle edits to existing hair to the addition of complex and challenging hair in images of clean-shaven subjects.
no code implementations • CVPR 2020 • Matheus Gadelha, Giorgio Gori, Duygu Ceylan, Radomir Mech, Nathan Carr, Tamy Boubekeur, Rui Wang, Subhransu Maji
We present a generative model to synthesize 3D shapes as sets of handles -- lightweight proxies that approximate the original 3D shape -- for applications in interactive editing, shape parsing, and building compact 3D representations.
1 code implementation • CVPR 2020 • Nenglun Chen, Lingjie Liu, Zhiming Cui, Runnan Chen, Duygu Ceylan, Changhe Tu, Wenping Wang
The 3D structure points produced by our method encode the shape structure intrinsically and exhibit semantic consistency across all the shape instances with similar structures.
no code implementations • ICCV 2019 • Christian Zimmermann, Duygu Ceylan, Jimei Yang, Bryan Russell, Max Argus, Thomas Brox
We show that methods trained on our dataset consistently perform well when tested on other datasets.
Ranked #24 on
3D Hand Pose Estimation
on FreiHAND
(PA-F@5mm metric)
3 code implementations • NeurIPS 2019 • Qiangeng Xu, Weiyue Wang, Duygu Ceylan, Radomir Mech, Ulrich Neumann
Reconstructing 3D shapes from single-view images has been a long-standing research problem.
Ranked #1 on
Single-View 3D Reconstruction
on ShapeNetCore
1 code implementation • CVPR 2019 • Weiyue Wang, Duygu Ceylan, Radomir Mech, Ulrich Neumann
Given such a source 3D model and a target which can be a 2D image, 3D model, or a point cloud acquired as a depth scan, we introduce 3DN, an end-to-end network that deforms the source model to resemble the target.
1 code implementation • ECCV 2018 • Amit Raj, Patsorn Sangkloy, Huiwen Chang, Jingwan Lu, Duygu Ceylan, James Hays
Garment transfer is a challenging task that requires (i) disentangling the features of the clothing from the body pose and shape and (ii) realistic synthesis of the garment texture on the new body.
Ranked #1 on
Virtual Try-on
on FashionIQ
(using extra training data)
no code implementations • 29 Jun 2018 • Tuanfeng Y. Wang, Duygu Ceylan, Jovan Popovic, Niloy J. Mitra
Designing real and virtual garments is becoming extremely demanding with rapidly changing fashion trends and increasing need for synthesizing realistic dressed digital humans for various applications.
Graphics
no code implementations • 20 Jun 2018 • Aron Monszpart, Paul Guerrero, Duygu Ceylan, Ersin Yumer, Niloy J. Mitra
A long-standing challenge in scene analysis is the recovery of scene arrangements under moderate to heavy occlusion, directly from monocular video.
1 code implementation • CVPR 2018 • Chen Liu, Jimei Yang, Duygu Ceylan, Ersin Yumer, Yasutaka Furukawa
The proposed end-to-end DNN learns to directly infer a set of plane parameters and corresponding plane segmentation masks from a single RGB image.
Ranked #2 on
Plane Instance Segmentation
on NYU Depth v2
1 code implementation • CVPR 2018 • Ruben Villegas, Jimei Yang, Duygu Ceylan, Honglak Lee
We propose a recurrent neural network architecture with a Forward Kinematics layer and cycle consistency based adversarial training objective for unsupervised motion retargetting.
2 code implementations • ECCV 2018 • Gül Varol, Duygu Ceylan, Bryan Russell, Jimei Yang, Ersin Yumer, Ivan Laptev, Cordelia Schmid
Human shape estimation is an important task for video editing, animation and fashion industry.
Ranked #3 on
3D Human Pose Estimation
on Surreal
(using extra training data)
no code implementations • ICCV 2017 • Ronald Yu, Shunsuke Saito, Haoxiang Li, Duygu Ceylan, Hao Li
To train such a network, we generate a massive dataset of synthetic faces with dense labels using renderings of a morphable face model with variations in pose, expressions, lighting, and occlusions.
2 code implementations • ICCV 2017 • Chuhang Zou, Ersin Yumer, Jimei Yang, Duygu Ceylan, Derek Hoiem
The success of various applications including robotics, digital content creation, and visualization demand a structured and abstract representation of the 3D world from limited sensor data.
no code implementations • ICCV 2017 • Guilin Liu, Duygu Ceylan, Ersin Yumer, Jimei Yang, Jyh-Ming Lien
We propose an end-to-end network architecture that replicates the forward image formation process to accomplish this task.
no code implementations • 14 Jun 2017 • Haibin Huang, Evangelos Kalogerakis, Siddhartha Chaudhuri, Duygu Ceylan, Vladimir G. Kim, Ersin Yumer
We present a new local descriptor for 3D shapes, directly applicable to a wide range of shape analysis problems such as point correspondences, semantic segmentation, affordance prediction, and shape-to-scan matching.
2 code implementations • CVPR 2017 • Eunbyung Park, Jimei Yang, Ersin Yumer, Duygu Ceylan, Alexander C. Berg
Instead of taking a 'blank slate' approach, we first explicitly infer the parts of the geometry visible both in the input and novel views and then re-cast the remaining synthesis problem as image completion.
no code implementations • 20 Apr 2016 • Guilin Liu, Chao Yang, Zimo Li, Duygu Ceylan, Qi-Xing Huang
Due to the abundance of 2D product images from the Internet, developing efficient and scalable algorithms to recover the missing depth information is central to many applications.
no code implementations • 11 Apr 2016 • Ruizhe Wang, Lingyu Wei, Etienne Vouga, Qi-Xing Huang, Duygu Ceylan, Gerard Medioni, Hao Li
We present an end-to-end system for reconstructing complete watertight and textured models of moving subjects such as clothed humans and animals, using only three or four handheld sensors.
no code implementations • CVPR 2016 • Lingyu Wei, Qi-Xing Huang, Duygu Ceylan, Etienne Vouga, Hao Li
We propose a deep learning approach for finding dense correspondences between 3D scans of people.