Search Results for author: Miguel A. Otaduy

Found 6 papers, 3 papers with code

SNUG: Self-Supervised Neural Dynamic Garments

1 code implementation CVPR 2022 Igor Santesteban, Miguel A. Otaduy, Dan Casas

We present a self-supervised method to learn dynamic 3D deformations of garments worn by parametric human bodies.

RGB2Hands: Real-Time Tracking of 3D Hand Interactions from Monocular RGB Video

no code implementations22 Jun 2021 Jiayi Wang, Franziska Mueller, Florian Bernard, Suzanne Sorli, Oleksandr Sotnychenko, Neng Qian, Miguel A. Otaduy, Dan Casas, Christian Theobalt

Moreover, we demonstrate that our approach offers previously unseen two-hand tracking performance from RGB, and quantitatively and qualitatively outperforms existing RGB-based methods that were not explicitly designed for two-hand interactions.

3D Reconstruction Sign Language Recognition

Self-Supervised Collision Handling via Generative 3D Garment Models for Virtual Try-On

1 code implementation CVPR 2021 Igor Santesteban, Nils Thuerey, Miguel A. Otaduy, Dan Casas

We propose a new generative model for 3D garment deformations that enables us to learn, for the first time, a data-driven method for virtual try-on that effectively addresses garment-body collisions.

Virtual Try-on

SoftSMPL: Data-driven Modeling of Nonlinear Soft-tissue Dynamics for Parametric Humans

no code implementations1 Apr 2020 Igor Santesteban, Elena Garces, Miguel A. Otaduy, Dan Casas

We present SoftSMPL, a learning-based method to model realistic soft-tissue dynamics as a function of body shape and motion.

Learning-Based Animation of Clothing for Virtual Try-On

1 code implementation17 Mar 2019 Igor Santesteban, Miguel A. Otaduy, Dan Casas

We propose a model that separates global garment fit, due to body shape, from local garment wrinkles, due to both pose dynamics and body shape.

Virtual Try-on

Cannot find the paper you are looking for? You can Submit a new open access paper.