1 code implementation • 20 Sep 2023 • Tianbao Xie, Siheng Zhao, Chen Henry Wu, Yitao Liu, Qian Luo, Victor Zhong, Yanchao Yang, Tao Yu
Unlike inverse RL and recent work that uses LLMs to write sparse reward codes, Text2Reward produces interpretable, free-form dense reward codes that cover a wide range of tasks, utilize existing packages, and allow iterative refinement with human feedback.
no code implementations • 4 Apr 2023 • Dong Lao, Zhengyang Hu, Francesco Locatello, Yanchao Yang, Stefano Soatto
We introduce a method to segment the visual field into independently moving regions, trained with no ground truth or supervision.
1 code implementation • CVPR 2023 • Xiaomeng Xu, Yanchao Yang, Kaichun Mo, Boxiao Pan, Li Yi, Leonidas Guibas
We propose a method that trains a neural radiance field (NeRF) to encode not only the appearance of the scene but also semantic correlations between scene points, regions, or entities -- aiming to capture their mutual co-variation patterns.
1 code implementation • CVPR 2023 • Bingfan Zhu, Yanchao Yang, Xulong Wang, Youyi Zheng, Leonidas Guibas
We propose VDN-NeRF, a method to train neural radiance fields (NeRFs) for better geometry under non-Lambertian surface and dynamic lighting conditions that cause significant variation in the radiance of a point when viewed from different angles.
no code implementations • ICCV 2023 • Jingen Jiang, Mingyang Zhao, Shiqing Xin, Yanchao Yang, Hanxiao Wang, Xiaohong Jia, Dong-Ming Yan
We propose a novel and efficient method for reconstructing manifold surfaces from point clouds.
no code implementations • ICCV 2023 • Boxiao Pan, Bokui Shen, Davis Rempe, Despoina Paschalidou, Kaichun Mo, Yanchao Yang, Leonidas J. Guibas
In this work, we introduce the challenging problem of predicting collisions in diverse environments from multi-view egocentric videos captured from body-mounted cameras.
no code implementations • 13 Jul 2022 • Yang Zheng, Tolga Birdal, Fei Xia, Yanchao Yang, Yueqi Duan, Leonidas J. Guibas
To this end, we propose: (i) a hierarchical localization system, where we leverage temporal information and (ii) a novel environment-aware image enhancement method to boost the robustness and accuracy.
no code implementations • 12 Jul 2022 • Colton Stearns, Davis Rempe, Jie Li, Rares Ambrus, Sergey Zakharov, Vitor Guizilini, Yanchao Yang, Leonidas J Guibas
In this work, we develop a holistic representation of traffic scenes that leverages both spatial and temporal information of the actors in the scene.
1 code implementation • 20 Apr 2022 • Yang Zheng, Yanchao Yang, Kaichun Mo, Jiaman Li, Tao Yu, Yebin Liu, C. Karen Liu, Leonidas J. Guibas
We perform an extensive study of the benefits of leveraging the eye gaze for ego-centric human motion prediction with various state-of-the-art architectures.
no code implementations • CVPR 2022 • Hanxiang Ren, Yanchao Yang, He Wang, Bokui Shen, Qingnan Fan, Youyi Zheng, C. Karen Liu, Leonidas J. Guibas
We describe a method to deal with performance drop in semantic segmentation caused by viewpoint changes within multi-camera systems, where temporally paired images are readily available, but the annotations may only be abundant for a few typical views.
1 code implementation • CVPR 2022 • Yuefan Shen, Yanchao Yang, Mi Yan, He Wang, Youyi Zheng, Leonidas Guibas
Here we propose a simple yet effective method for unsupervised domain adaptation on point clouds by employing a self-supervised task of learning geometry-aware implicits, which plays two critical roles in one shot.
no code implementations • ICLR 2022 • Chuanyu Pan, Yanchao Yang, Kaichun Mo, Yueqi Duan, Leonidas Guibas
We perform an extensive study of the key features of the proposed framework and analyze the characteristics of the learned representations.
no code implementations • ICLR 2022 • Qi Li, Kaichun Mo, Yanchao Yang, Hang Zhao, Leonidas Guibas
While most works focus on single-object or agent-object visual functionality and affordances, our work proposes to study a new kind of visual relationship that is also important to perceive and model -- inter-object functional relationships (e. g., a switch on the wall turns on or off the light, a remote control operates the TV).
1 code implementation • 29 Jul 2021 • Yanchao Yang, Hanxiang Ren, He Wang, Bokui Shen, Qingnan Fan, Youyi Zheng, C. Karen Liu, Leonidas Guibas
Furthermore, to resolve ambiguities in converting the semantic images to semantic labels, we treat the view transformation network as a functional representation of an unknown mapping implied by the color images and propose functional label hallucination to generate pseudo-labels in the target domain.
2 code implementations • 27 Jul 2021 • Yuefan Shen, Yanchao Yang, Youyi Zheng, C. Karen Liu, Leonidas Guibas
We describe a method for unpaired realistic depth synthesis that learns diverse variations from the real-world depth scans and ensures geometric consistency between the synthetic and synthesized depth.
no code implementations • CVPR 2021 • Xinzhu Bei, Yanchao Yang, Stefano Soatto
The appearance of the scene is warped from past frames using the predicted motion in co-visible regions; dis-occluded regions are synthesized with content-aware inpainting utilizing the predicted scene layout.
no code implementations • CVPR 2021 • Yanchao Yang, Brian Lai, Stefano Soatto
Then, it uses the segments to learn object models that can be used for detection in a static image.
3 code implementations • CVPR 2020 • Yanchao Yang, Stefano Soatto
We describe a simple method for unsupervised domain adaptation, whereby the discrepancy between the source and target distributions is reduced by swapping the low-frequency spectrum of one with the other.
Ranked #3 on
Domain Adaptation
on Panoptic SYNTHIA-to-Mapillary
1 code implementation • CVPR 2020 • Yanchao Yang, Yutong Chen, Stefano Soatto
We describe a method to train a generative model with latent factors that are (approximately) independent and localized.
1 code implementation • CVPR 2020 • Yanchao Yang, Dong Lao, Ganesh Sundaramoorthi, Stefano Soatto
We introduce two criteria to regularize the optimization involved in learning a classifier in a domain where no annotated data are available, leveraging annotated data in a different domain, a problem known as unsupervised domain adaptation.
no code implementations • CVPR 2019 • Yanchao Yang, Alex Wong, Stefano Soatto
We present a deep learning system to infer the posterior distribution of a dense depth map associated with an image, by exploiting sparse range measurements, for instance from a lidar.
Ranked #5 on
Depth Completion
on VOID
1 code implementation • CVPR 2019 • Yanchao Yang, Antonio Loquercio, Davide Scaramuzza, Stefano Soatto
We propose an adversarial contextual model for detecting moving objects in images.
1 code implementation • ECCV 2018 • Yanchao Yang, Stefano Soatto
On the other hand, fully supervised methods learn the regularity in the annotated data, without explicit regularization and with the risk of overfitting.
no code implementations • CVPR 2017 • Yanchao Yang, Stefano Soatto
We introduce a method to compute optical flow at multiple scales of motion, without resorting to multi- resolution or combinatorial methods.
no code implementations • ICCV 2015 • Yanchao Yang, Ganesh Sundaramoorthi, Stefano Soatto
We propose a method to detect disocclusion in video sequences of three-dimensional scenes and to partition the disoccluded regions into objects, defined by coherent deformation corresponding to surfaces in the scene.
no code implementations • CVPR 2015 • Yanchao Yang, Zhaojin Lu, Ganesh Sundaramoorthi
We present a new approach to wide baseline matching.
no code implementations • 21 Aug 2012 • Yanchao Yang, Ganesh Sundaramoorthi
In cases of 3D object motion and viewpoint change, self-occlusions and dis-occlusions of the object are prominent, and current methods employing joint shape and appearance models are unable to adapt to new shape and appearance information, leading to inaccurate shape detection.