no code implementations • 22 Apr 2024 • Mana Masuda, Jinhyung Park, Shun Iwase, Rawal Khirodkar, Kris Kitani
While recent advancements in animatable human rendering have achieved remarkable results, they require test-time optimization for each subject which can be a significant limitation for real-world applications.
no code implementations • 28 Jan 2024 • Yu-Jhe Li, Yan Xu, Rawal Khirodkar, Jinhyung Park, Kris Kitani
In order to evaluate our proposed pipeline, we collect three video sets of RGBD videos recorded from multiple sparse-view depth cameras and ground truth 3D poses are manually annotated.
no code implementations • CVPR 2024 • Jinhyung Park, Yu-Jhe Li, Kris Kitani
While recent depth completion methods have achieved remarkable results filling in relatively dense depth maps (e. g. projected 64-line LiDAR on KITTI or 500 sampled points on NYUv2) with RGB guidance their performance on very sparse input (e. g. 4-line LiDAR or 32 depth point measurements) is unverified.
no code implementations • 7 Apr 2023 • Sang-Bin Jeon, Jaeho Jung, Jinhyung Park, In-Kwon Lee
In order to serve better VR experiences to users, existing predictive methods of Redirected Walking (RDW) exploit future information to reduce the number of reset occurrences.
1 code implementation • CVPR 2023 • Yu-Jhe Li, Shawn Hunt, Jinhyung Park, Matthew O’Toole, Kris Kitani
We also propose a hybrid super-resolution model (Hybrid-SR) combining our ADC-SR with a standard RAD super-resolution model, and show that performance can be improved by a large margin.
1 code implementation • 5 Oct 2022 • Jinhyung Park, Chenfeng Xu, Shijia Yang, Kurt Keutzer, Kris Kitani, Masayoshi Tomizuka, Wei Zhan
While recent camera-only 3D detection methods leverage multiple timesteps, the limited history they use significantly hampers the extent to which temporal fusion can improve object perception.
Ranked #1 on Robust Camera Only 3D Object Detection on nuScenes-C
1 code implementation • 17 Mar 2022 • Jinhyung Park, Chenfeng Xu, Yiyang Zhou, Masayoshi Tomizuka, Wei Zhan
While numerous 3D detection works leverage the complementary relationship between RGB images and point clouds, developments in the broader framework of semi-supervised object recognition remain uninfluenced by multi-modal fusion.
no code implementations • CVPR 2022 • Yu-Jhe Li, Jinhyung Park, Matthew O'Toole, Kris Kitani
To mitigate this problem, we propose the Self-Training Multimodal Vehicle Detection Network (ST-MVDNet) which leverages a Teacher-Student mutual learning framework and a simulated sensor noise model used in strong data augmentation for Lidar and Radar.
1 code implementation • 7 Dec 2021 • Jinhyung Park, Dohae Lee, In-Kwon Lee
This precondition restrains the model from generalizing to real-world data, which is considered to be a sequence of unordered point sets.
1 code implementation • 8 Jul 2021 • Jinhyung Park, Xinshuo Weng, Yunze Man, Kris Kitani
To provide a more integrated approach, we propose a novel Multi-Modality Task Cascade network (MTC-RCNN) that leverages 3D box proposals to improve 2D segmentation predictions, which are then used to further refine the 3D boxes.