no code implementations • 12 Jun 2025 • Justin Kerr, Kush Hari, Ethan Weber, Chung Min Kim, Brent Yi, Tyler Bonnen, Ken Goldberg, Angjoo Kanazawa
In this way, hand-eye coordination emerges as the eye looks towards regions which allow the hand to complete the task.
no code implementations • 6 May 2025 • Arthur Allshire, Hongsuk Choi, Junyi Zhang, David McAllister, Anthony Zhang, Chung Min Kim, Trevor Darrell, Pieter Abbeel, Jitendra Malik, Angjoo Kanazawa
How can we teach humanoids to climb staircases and sit on chairs using the surrounding environment context?
no code implementations • 17 Apr 2025 • Haiwen Feng, Junyi Zhang, Qianqian Wang, Yufei Ye, Pengcheng Yu, Michael J. Black, Trevor Darrell, Angjoo Kanazawa
Dynamic 3D reconstruction and point tracking in videos are typically treated as separate tasks, despite their deep connection.
no code implementations • CVPR 2025 • Nan Huang, Wenzhao Zheng, Chenfeng Xu, Kurt Keutzer, Shanghang Zhang, Angjoo Kanazawa, Qianqian Wang
Moving object segmentation is a crucial task for achieving a high-level understanding of visual scenes and has numerous downstream applications.
no code implementations • 7 Feb 2025 • Ethan Weber, Norman Müller, Yash Kant, Vasu Agrawal, Michael Zollhöfer, Angjoo Kanazawa, Christian Richardt
Our solution is to train a generative model that can consume a large context of input frames while generating unknown target views and recovering image poses when desired.
no code implementations • CVPR 2025 • Qianqian Wang, Yifei Zhang, Aleksander Holynski, Alexei A. Efros, Angjoo Kanazawa
We present a unified framework capable of solving a broad range of 3D tasks.
no code implementations • CVPR 2025 • David McAllister, Matthew Tancik, Jiaming Song, Angjoo Kanazawa
We propose Decentralized Diffusion Models, a scalable framework for distributing diffusion model training across independent clusters or datacenters by eliminating the dependence on a centralized, high-bandwidth networking fabric.
no code implementations • CVPR 2025 • Zhengqi Li, Richard Tucker, Forrester Cole, Qianqian Wang, Linyi Jin, Vickie Ye, Angjoo Kanazawa, Aleksander Holynski, Noah Snavely
We present a system that allows for accurate, fast, and robust estimation of camera parameters and depth maps from casual monocular videos of dynamic scenes.
1 code implementation • CVPR 2025 • Lea Müller, Hongsuk Choi, Anthony Zhang, Brent Yi, Jitendra Malik, Angjoo Kanazawa
We present "Humans and Structure from Motion" (HSfM), a method for jointly reconstructing multiple human meshes, scene point clouds, and camera parameters in a metric world coordinate system from a sparse set of uncalibrated multi-view images featuring people.
1 code implementation • 5 Dec 2024 • Zhengqi Li, Richard Tucker, Forrester Cole, Qianqian Wang, Linyi Jin, Vickie Ye, Angjoo Kanazawa, Aleksander Holynski, Noah Snavely
We present a system that allows for accurate, fast, and robust estimation of camera parameters and depth maps from casual monocular videos of dynamic scenes.
no code implementations • 31 Oct 2024 • Zhuoyang Pan, Angjoo Kanazawa, Hang Gao
SOAR leverages structural normal prior and generative diffusion prior to address such an ill-posed reconstruction problem.
no code implementations • 21 Oct 2024 • Gengshan Yang, Andrea Bajcsy, Shunsuke Saito, Angjoo Kanazawa
We present Agent-to-Sim (ATS), a framework for learning interactive behavior models of 3D agents from casual longitudinal video collections.
no code implementations • CVPR 2025 • Brent Yi, Vickie Ye, Maya Zheng, Yunqi Li, Lea Müller, Georgios Pavlakos, Yi Ma, Jitendra Malik, Angjoo Kanazawa
We present EgoAllo, a system for human motion estimation from a head-mounted device.
1 code implementation • 26 Sep 2024 • Justin Kerr, Chung Min Kim, Mingxuan Wu, Brent Yi, Qianqian Wang, Ken Goldberg, Angjoo Kanazawa
This analysis-by-synthesis approach uses part-centric feature fields in an iterative optimization which enables the use of geometric regularizers to recover 3D motions from only a single video.
1 code implementation • 10 Sep 2024 • Vickie Ye, RuiLong Li, Justin Kerr, Matias Turkulainen, Brent Yi, Zhuoyang Pan, Otto Seiskari, Jianbo Ye, Jeffrey Hu, Matthew Tancik, Angjoo Kanazawa
gsplat is an open-source library designed for training and developing Gaussian Splatting methods.
no code implementations • 6 Sep 2024 • Vongani Maluleke, Lea Müller, Jathushan Rajasegaran, Georgios Pavlakos, Shiry Ginosar, Angjoo Kanazawa, Jitendra Malik
Our contributions are a demonstration of the advantages of socially conditioned future motion prediction and an in-the-wild, couple dance video dataset to enable future research in this direction.
no code implementations • 18 Jul 2024 • Qianqian Wang, Vickie Ye, Hang Gao, Jake Austin, Zhengqi Li, Angjoo Kanazawa
Monocular dynamic reconstruction is a challenging and long-standing vision problem due to the highly ill-posed nature of the task.
no code implementations • 17 Jul 2024 • Congrong Xu, Justin Kerr, Angjoo Kanazawa
Novel view synthesis from unconstrained in-the-wild image collections remains a significant yet challenging task due to photometric variations and transient occluders that complicate accurate scene reconstruction.
no code implementations • 13 Jun 2024 • David McAllister, Songwei Ge, Jia-Bin Huang, David W. Jacobs, Alexei A. Efros, Aleksander Holynski, Angjoo Kanazawa
We compare our method to existing approaches for score distillation sampling and show that it can produce high-frequency details with realistic colors.
no code implementations • 16 May 2024 • Ethan Weber, Riley Peterlinz, Rohan Mathur, Frederik Warburg, Alexei A. Efros, Angjoo Kanazawa
We recover the underlying 3D structure from images of cartoons and anime depicting the same scene.
no code implementations • 9 May 2024 • Yash Khandelwal, Mayur Arvind, Sriram Kumar, Ashish Gupta, Sachin Kumar Danisetty, Piyush Bagad, Anish Madan, Mayank Lunayach, Aditya Annavajjala, Abhishek Maiti, Sansiddh Jain, Aman Dalmia, Namrata Deka, Jerome White, Jigar Doshi, Angjoo Kanazawa, Rahul Panicker, Alpan Raval, Srinivas Rana, Makarand Tapaswi
Our goal is to equip health workers and public health systems with a solution for contactless newborn anthropometry in the community.
no code implementations • 24 Apr 2024 • RuiLong Li, Sanja Fidler, Angjoo Kanazawa, Francis Williams
We present NeRF-XL, a principled method for distributing Neural Radiance Fields (NeRFs) across multiple GPUs, thus enabling the training and rendering of NeRFs with an arbitrarily large capacity.
no code implementations • 7 Apr 2024 • Chiara Plizzari, Shubham Goel, Toby Perrett, Jacob Chalk, Angjoo Kanazawa, Dima Damen
In this paper, we aim to mimic this spatial cognition ability.
1 code implementation • 4 Apr 2024 • Xinyang Han, Zelin Gao, Angjoo Kanazawa, Shubham Goel, Yossi Gandelsman
Inspired by this behavior, we introduce SAP3D, a system for 3D reconstruction and novel view synthesis from an arbitrary number of unposed images.
1 code implementation • CVPR 2024 • Chung Min Kim, Mingxuan Wu, Justin Kerr, Ken Goldberg, Matthew Tancik, Angjoo Kanazawa
We optimize this field from a set of 2D masks provided by Segment Anything (SAM) in a way that respects coarse-to-fine hierarchy, using scale to consistently fuse conflicting masks from different viewpoints.
1 code implementation • CVPR 2024 • Evonne Ng, Javier Romero, Timur Bagautdinov, Shaojie Bai, Trevor Darrell, Angjoo Kanazawa, Alexander Richard
We present a framework for generating full-bodied photorealistic avatars that gesture according to the conversational dynamics of a dyadic interaction.
1 code implementation • CVPR 2024 • Xinyang Han, Zelin Gao, Angjoo Kanazawa, Shubham Goel, Yossi Gandelsman
Inspired by this behavior we introduce SAP3D a system for 3D reconstruction and novel view synthesis from an arbitrary number of unposed images.
1 code implementation • CVPR 2024 • Georgios Pavlakos, Dandan Shan, Ilija Radosavovic, Angjoo Kanazawa, David Fouhey, Jitendra Malik
The key to HaMeR's success lies in scaling up both the data used for training and the capacity of the deep network for hand reconstruction.
no code implementations • CVPR 2024 • Ethan Weber, Aleksander Hołyński, Varun Jampani, Saurabh Saxena, Noah Snavely, Abhishek Kar, Angjoo Kanazawa
In contrast to related works, we focus on completing scenes rather than deleting foreground objects, and our approach does not require tight 2D object masks or text.
1 code implementation • 4 Dec 2023 • Vickie Ye, Angjoo Kanazawa
This report provides the mathematical details of the gsplat library, a modular toolbox for efficient differentiable Gaussian splatting, as proposed by Kerbl et al.
no code implementations • 11 Oct 2023 • Ryan Po, Wang Yifan, Vladislav Golyanik, Kfir Aberman, Jonathan T. Barron, Amit H. Bermano, Eric Ryan Chan, Tali Dekel, Aleksander Holynski, Angjoo Kanazawa, C. Karen Liu, Lingjie Liu, Ben Mildenhall, Matthias Nießner, Björn Ommer, Christian Theobalt, Peter Wonka, Gordon Wetzstein
The field of visual computing is rapidly advancing due to the emergence of generative artificial intelligence (AI), which unlocks unprecedented capabilities for the generation, editing, and reconstruction of images, videos, and 3D scenes.
no code implementations • 14 Sep 2023 • Adam Rashid, Satvik Sharma, Chung Min Kim, Justin Kerr, Lawrence Chen, Angjoo Kanazawa, Ken Goldberg
Instead, we propose LERF-TOGO, Language Embedded Radiance Fields for Task-Oriented Grasping of Objects, which uses vision-language models zero-shot to output a grasp distribution over an object given a natural language query.
no code implementations • ICCV 2023 • Evonne Ng, Sanjay Subramanian, Dan Klein, Angjoo Kanazawa, Trevor Darrell, Shiry Ginosar
We present a framework for generating appropriate facial responses from a listener in dyadic social interactions based on the speaker's words.
no code implementations • NeurIPS 2023 • Tom Monnier, Jake Austin, Angjoo Kanazawa, Alexei A. Efros, Mathieu Aubry
We compare our approach to the state of the art on diverse scenes from DTU, and demonstrate its robustness on real-life captures from BlendedMVS and Nerfstudio.
1 code implementation • CVPR 2024 • Lea Müller, Vickie Ye, Georgios Pavlakos, Michael Black, Angjoo Kanazawa
To address this, we present a novel approach that learns a prior over the 3D proxemics two people in close social interaction and demonstrate its use for single-view 3D reconstruction.
1 code implementation • ICCV 2023 • Shubham Goel, Georgios Pavlakos, Jathushan Rajasegaran, Angjoo Kanazawa, Jitendra Malik
To analyze video, we use 3D reconstructions from HMR 2. 0 as input to a tracking system that operates in 3D.
Ranked #3 on
Pose Tracking
on PoseTrack2018
no code implementations • ICCV 2023 • RuiLong Li, Hang Gao, Matthew Tancik, Angjoo Kanazawa
Optimizing and rendering Neural Radiance Fields is computationally expensive due to the vast number of samples required by volume rendering.
1 code implementation • ICCV 2023 • Frederik Warburg, Ethan Weber, Matthew Tancik, Aleksander Holynski, Angjoo Kanazawa
Casually captured Neural Radiance Fields (NeRFs) suffer from artifacts such as floaters or flawed geometry when rendered outside the camera trajectory.
no code implementations • 4 Apr 2023 • Aymen Mir, Xavier Puig, Angjoo Kanazawa, Gerard Pons-Moll
We decompose the continual motion synthesis problem into walking along paths and transitioning in and out of the actions specified by the keypoints, which enables long generation of motions that satisfy scene constraints without explicitly incorporating scene information.
1 code implementation • CVPR 2023 • Jathushan Rajasegaran, Georgios Pavlakos, Angjoo Kanazawa, Christoph Feichtenhofer, Jitendra Malik
Subsequently, we propose a Lagrangian Action Recognition model by fusing 3D pose and contextualized appearance over tracklets.
Ranked #1 on
Action Recognition
on AVA v2.2
(using extra training data)
1 code implementation • ICCV 2023 • Ayaan Haque, Matthew Tancik, Alexei A. Efros, Aleksander Holynski, Angjoo Kanazawa
We propose a method for editing NeRF scenes with text-instructions.
5 code implementations • ICCV 2023 • Justin Kerr, Chung Min Kim, Ken Goldberg, Angjoo Kanazawa, Matthew Tancik
Humans describe the physical world using natural language to refer to specific 3D locations based on a vast range of properties: visual appearance, semantics, abstract associations, or actionable affordances.
1 code implementation • CVPR 2023 • Vickie Ye, Georgios Pavlakos, Jitendra Malik, Angjoo Kanazawa
Our method robustly recovers the global 3D trajectories of people in challenging in-the-wild videos, such as PoseTrack.
2 code implementations • 8 Feb 2023 • Matthew Tancik, Ethan Weber, Evonne Ng, RuiLong Li, Brent Yi, Justin Kerr, Terrance Wang, Alexander Kristoffersen, Jake Austin, Kamyar Salahi, Abhik Ahuja, David McAllister, Angjoo Kanazawa
Neural Radiance Fields (NeRF) are a rapidly growing area of research with wide-ranging applications in computer vision, graphics, robotics, and more.
Ranked #7 on
Novel View Synthesis
on RefRef
2 code implementations • CVPR 2023 • Sara Fridovich-Keil, Giacomo Meanti, Frederik Warburg, Benjamin Recht, Angjoo Kanazawa
We introduce k-planes, a white-box model for radiance fields in arbitrary dimensions.
Ranked #2 on
Novel View Synthesis
on LLFF
1 code implementation • 24 Oct 2022 • Hang Gao, RuiLong Li, Shubham Tulsiani, Bryan Russell, Angjoo Kanazawa
We study the recent progress on dynamic view synthesis (DVS) from monocular video.
1 code implementation • 10 Oct 2022 • RuiLong Li, Matthew Tancik, Angjoo Kanazawa
We propose NerfAcc, a toolbox for efficient volumetric rendering of radiance fields.
no code implementations • 6 Sep 2022 • Vongani H. Maluleke, Neerja Thakkar, Tim Brooks, Ethan Weber, Trevor Darrell, Alexei A. Efros, Angjoo Kanazawa, Devin Guillory
In this work, we study how the performance and evaluation of generative image models are impacted by the racial composition of their training datasets.
no code implementations • 28 Jul 2022 • Georgios Pavlakos, Ethan Weber, Matthew Tancik, Angjoo Kanazawa
TV shows depict a wide variety of human behaviors and have been studied extensively for their potential to be a rich source of data for many applications.
1 code implementation • 22 Jul 2022 • Zhengqi Li, Qianqian Wang, Noah Snavely, Angjoo Kanazawa
We present a method for learning to generate unbounded flythrough videos of natural scenes starting from a single view, where this capability is learned from a collection of single photographs, without requiring camera poses or even multiple views of each scene.
1 code implementation • 21 Jun 2022 • Zhenzhen Weng, Kuan-Chieh Wang, Angjoo Kanazawa, Serena Yeung
The ability to perceive 3D human bodies from a single image has a multitude of applications ranging from entertainment and robotics to neuroscience and healthcare.
1 code implementation • 17 Jun 2022 • RuiLong Li, Julian Tanke, Minh Vo, Michael Zollhofer, Jurgen Gall, Angjoo Kanazawa, Christoph Lassner
Since TAVA does not require a body template, it is applicable to humans as well as other creatures such as animals.
no code implementations • CVPR 2022 • Evonne Ng, Hanbyul Joo, Liwen Hu, Hao Li, Trevor Darrell, Angjoo Kanazawa, Shiry Ginosar
We present a framework for modeling interactional communication in dyadic conversations: given multimodal inputs of a speaker, we autoregressively output multiple possibilities of corresponding listener motion.
1 code implementation • CVPR 2022 • Vickie Ye, Zhengqi Li, Richard Tucker, Angjoo Kanazawa, Noah Snavely
We describe a method to extract persistent elements of a dynamic scene from an input video.
no code implementations • CVPR 2022 • Jathushan Rajasegaran, Georgios Pavlakos, Angjoo Kanazawa, Jitendra Malik
For a future frame, we compute the similarity between the predicted state of a tracklet and the single frame observations in a probabilistic manner.
4 code implementations • CVPR 2022 • Alex Yu, Sara Fridovich-Keil, Matthew Tancik, Qinhong Chen, Benjamin Recht, Angjoo Kanazawa
We introduce Plenoxels (plenoptic voxels), a system for photorealistic view synthesis.
no code implementations • 8 Dec 2021 • Jathushan Rajasegaran, Georgios Pavlakos, Angjoo Kanazawa, Jitendra Malik
For a future frame, we compute the similarity between the predicted state of a tracklet and the single frame observations in a probabilistic manner.
1 code implementation • NeurIPS 2021 • Jathushan Rajasegaran, Georgios Pavlakos, Angjoo Kanazawa, Jitendra Malik
We find that 3D representations are more effective than 2D representations for tracking in these settings, and we obtain state-of-the-art performance.
no code implementations • ICLR 2022 • Shizhan Zhu, Sayna Ebrahimi, Angjoo Kanazawa, Trevor Darrell
Existing approaches for single object reconstruction impose supervision signals based on the loss of the signed distance value from all locations in a scene, posing difficulties when extending to real-world scenarios.
no code implementations • CVPR 2021 • Tomas Jakab, Richard Tucker, Ameesh Makadia, Jiajun Wu, Noah Snavely, Angjoo Kanazawa
We cast this as the problem of aligning a source 3D object to a target 3D object from the same object category.
1 code implementation • CVPR 2021 • Shangzhe Wu, Ameesh Makadia, Jiajun Wu, Noah Snavely, Richard Tucker, Angjoo Kanazawa
Recent works have shown exciting results in unsupervised image de-rendering -- learning to decompose 3D shape, appearance, and lighting from single-image collections without explicit supervision.
4 code implementations • 5 Apr 2021 • Xue Bin Peng, Ze Ma, Pieter Abbeel, Sergey Levine, Angjoo Kanazawa
Our system produces high-quality motions that are comparable to those achieved by state-of-the-art tracking-based techniques, while also being able to easily accommodate large datasets of unstructured motion clips.
5 code implementations • ICCV 2021 • Alex Yu, RuiLong Li, Matthew Tancik, Hao Li, Ren Ng, Angjoo Kanazawa
We introduce a method to render Neural Radiance Fields (NeRFs) in real time using PlenOctrees, an octree-based 3D representation which supports view-dependent effects.
1 code implementation • ICCV 2021 • RuiLong Li, Shan Yang, David A. Ross, Angjoo Kanazawa
We present AIST++, a new multi-modal dataset of 3D dance motion and music, along with FACT, a Full-Attention Cross-modal Transformer network for generating 3D dance motion conditioned on music.
Ranked #2 on
Motion Synthesis
on BRACE
1 code implementation • CVPR 2022 • Georgios Pavlakos, Jitendra Malik, Angjoo Kanazawa
The tools we develop open the door to processing and analyzing in 3D content from a large library of edited media, which could be helpful for many downstream applications.
1 code implementation • ICCV 2021 • Andrew Liu, Richard Tucker, Varun Jampani, Ameesh Makadia, Noah Snavely, Angjoo Kanazawa
We introduce the problem of perpetual view generation - long-range generation of novel views corresponding to an arbitrarily long camera trajectory given a single image.
no code implementations • ICCV 2021 • Zhe Cao, Ilija Radosavovic, Angjoo Kanazawa, Jitendra Malik
In this work we explore reconstructing hand-object interactions in the wild.
2 code implementations • CVPR 2021 • Alex Yu, Vickie Ye, Matthew Tancik, Angjoo Kanazawa
This allows the network to be trained across multiple scenes to learn a scene prior, enabling it to perform novel view synthesis in a feed-forward manner from a sparse set of views (as few as one).
Ranked #2 on
Generalizable Novel View Synthesis
on NERDS 360
2 code implementations • ECCV 2020 • Jason Y. Zhang, Sam Pepose, Hanbyul Joo, Deva Ramanan, Jitendra Malik, Angjoo Kanazawa
We present a method that infers spatial arrangements and shapes of humans and objects in a globally consistent 3D scene, all from a single image in-the-wild captured in an uncontrolled environment.
Ranked #3 on
3D Object Reconstruction
on BEHAVE
no code implementations • ECCV 2020 • Shubham Goel, Angjoo Kanazawa, Jitendra Malik
We present a learning framework that learns to recover the 3D shape, pose and texture from a single image, trained on an image collection without any ground truth 3D shape, multi-view, camera viewpoints or keypoint supervision.
2 code implementations • NeurIPS 2020 • Jake Levinson, Carlos Esteves, Kefan Chen, Noah Snavely, Angjoo Kanazawa, Afshin Rostamizadeh, Ameesh Makadia
Symmetric orthogonalization via SVD, and closely related procedures, are well-known techniques for projecting matrices onto $O(n)$ or $SO(n)$.
1 code implementation • ICCV 2019 • Silvia Zuffi, Angjoo Kanazawa, Tanya Berger-Wolf, Michael J. Black
In contrast to research on human pose, shape and texture estimation, training data for endangered species is limited, the animals are in complex natural scenes with occlusion, they are naturally camouflaged, travel in herds, and look similar to each other.
1 code implementation • ICCV 2019 • Jason Y. Zhang, Panna Felsen, Angjoo Kanazawa, Jitendra Malik
In this work, we present perhaps the first approach for predicting a future 3D mesh model sequence of a person from past video input.
1 code implementation • ICCV 2019 • Shunsuke Saito, Zeng Huang, Ryota Natsume, Shigeo Morishima, Angjoo Kanazawa, Hao Li
We introduce Pixel-aligned Implicit Function (PIFu), a highly effective implicit representation that locally aligns pixels of 2D images with the global context of their corresponding 3D object.
Ranked #1 on
3D Object Reconstruction
on RenderPeople
1 code implementation • CVPR 2019 • Angjoo Kanazawa, Jason Y. Zhang, Panna Felsen, Jitendra Malik
We present a framework that can similarly learn a representation of 3D dynamics of humans from video via a simple but effective temporal encoding of image features.
Ranked #17 on
3D Human Pose Estimation
on 3DPW
(Acceleration Error metric)
1 code implementation • 8 Oct 2018 • Xue Bin Peng, Angjoo Kanazawa, Jitendra Malik, Pieter Abbeel, Sergey Levine
In this paper, we propose a method that enables physically simulated characters to learn skills from videos (SFV).
5 code implementations • ICLR 2019 • Xue Bin Peng, Angjoo Kanazawa, Sam Toyer, Pieter Abbeel, Sergey Levine
By enforcing a constraint on the mutual information between the observations and the discriminator's internal representation, we can effectively modulate the discriminator's accuracy and maintain useful and informative gradients.
no code implementations • CVPR 2018 • Soumyadip Sengupta, Angjoo Kanazawa, Carlos D. Castillo, David W. Jacobs
SfSNet learns from a mixture of labeled synthetic and unlabeled real world images.
no code implementations • CVPR 2018 • Silvia Zuffi, Angjoo Kanazawa, Michael J. Black
Animals are widespread in nature and the analysis of their shape and motion is important in many fields and industries.
no code implementations • ECCV 2018 • Angjoo Kanazawa, Shubham Tulsiani, Alexei A. Efros, Jitendra Malik
The shape is represented as a deformable 3D mesh model of an object category where a shape is parameterized by a learned mean shape and per-instance predicted deformation.
10 code implementations • CVPR 2018 • Angjoo Kanazawa, Michael J. Black, David W. Jacobs, Jitendra Malik
The main objective is to minimize the reprojection loss of keypoints, which allow our model to be trained using images in-the-wild that only have ground truth 2D annotations.
Ranked #1 on
Weakly-supervised 3D Human Pose Estimation
on Human3.6M
(3D Annotations metric)
1 code implementation • CVPR 2018 • Soumyadip Sengupta, Angjoo Kanazawa, Carlos D. Castillo, David Jacobs
SfSNet learns from a mixture of labeled synthetic and unlabeled real world images.
no code implementations • 24 Jul 2017 • Yinghao Huang, Federica Bogo, Christoph Lassner, Angjoo Kanazawa, Peter V. Gehler, Ijaz Akhter, Michael J. Black
Existing marker-less motion capture methods often assume known backgrounds, static cameras, and sequence specific motion priors, which narrows its application scenarios.
no code implementations • CVPR 2017 • Silvia Zuffi, Angjoo Kanazawa, David Jacobs, Michael J. Black
The best human body models are learned from thousands of 3D scans of people in specific poses, which is infeasible with live animals.
2 code implementations • 27 Jul 2016 • Federica Bogo, Angjoo Kanazawa, Christoph Lassner, Peter Gehler, Javier Romero, Michael J. Black
We then fit (top-down) a recently published statistical body shape model, called SMPL, to the 2D joints.
Ranked #31 on
3D Human Pose Estimation
on HumanEva-I
no code implementations • CVPR 2016 • Angjoo Kanazawa, David W. Jacobs, Manmohan Chandraker
This is in contrast to prior works that require part annotations, since matching objects across class and pose variations is challenging with appearance features alone.
no code implementations • 28 Jul 2015 • Angjoo Kanazawa, Shahar Kovalsky, Ronen Basri, David W. Jacobs
In this paper, we show that such information can be learned from user-clicked 2D images and a template 3D model of the target animal.
no code implementations • 16 Dec 2014 • Angjoo Kanazawa, Abhishek Sharma, David Jacobs
We show on a modified MNIST dataset that when faced with scale variation, building in scale-invariance allows ConvNets to learn more discriminative features with reduced chances of over-fitting.