no code implementations • 18 Nov 2024 • Dongseok Shim, Yichun Shi, Kejie Li, H. Jin Kim, Peng Wang
Recent advancements in text-to-3D generation, building on the success of high-performance text-to-image generative models, have made it possible to create imaginative and richly textured 3D objects from textual descriptions.
no code implementations • 9 Jul 2024 • Xiaoding Yuan, Shitao Tang, Kejie Li, Alan Yuille, Peng Wang
This paper introduces Camera-free Diffusion (CamFreeDiff) model for 360-degree image outpainting from a single camera-free image and text description.
no code implementations • 26 Apr 2024 • SeungWook Kim, Yichun Shi, Kejie Li, Minsu Cho, Peng Wang
Using image as prompts for 3D generation demonstrate particularly strong performances compared to using text prompts alone, for images provide a more intuitive guidance for the 3D generation process.
no code implementations • CVPR 2024 • SeungWook Kim, Kejie Li, Xueqing Deng, Yichun Shi, Minsu Cho, Peng Wang
Leveraging multi-view diffusion models as priors for 3D optimization have alleviated the problem of 3D consistency, e. g., the Janus face problem or the content drift problem, in zero-shot text-to-3D models.
no code implementations • 4 Oct 2023 • Jianglong Ye, Peng Wang, Kejie Li, Yichun Shi, Heng Wang
Specifically, we decompose the NVS task into two stages: (i) transforming observed regions to a novel view, and (ii) hallucinating unseen regions.
4 code implementations • 31 Aug 2023 • Yichun Shi, Peng Wang, Jianglong Ye, Mai Long, Kejie Li, Xiao Yang
We introduce MVDream, a diffusion model that is able to generate consistent multi-view images from a given text prompt.
1 code implementation • ICCV 2023 • Qianyi Wu, Kaisiyuan Wang, Kejie Li, Jianmin Zheng, Jianfei Cai
Unlike traditional multi-view stereo approaches, the neural implicit surface-based methods leverage neural networks to represent 3D scenes as signed distance functions (SDFs).
no code implementations • 27 Jul 2023 • Sandika Biswas, Kejie Li, Biplab Banerjee, Subhasis Chaudhuri, Hamid Rezatofighi
This paper proposes using an implicit feature representation of the scene elements to distinguish a physically plausible alignment of humans and objects from an implausible one.
no code implementations • 22 May 2023 • Theo W. Costain, Kejie Li, Victor A. Prisacariu
Prior works have demonstrated that implicit representations trained only for reconstruction tasks typically generate encodings that are not useful for semantic tasks.
1 code implementation • CVPR 2024 • Shuai Chen, Yash Bhalgat, Xinghui Li, Jiawang Bian, Kejie Li, ZiRui Wang, Victor Adrian Prisacariu
To enhance the robustness of our model, we introduce a feature fusion module and a progressive training strategy.
1 code implementation • CVPR 2023 • Kejie Li, Jia-Wang Bian, Robert Castle, Philip H. S. Torr, Victor Adrian Prisacariu
The distinct data modality offered by high-resolution RGB images and low-resolution depth maps captured on a mobile device, when combined with precise 3D geometry annotations, presents a unique opportunity for future research on high-fidelity 3D reconstruction.
1 code implementation • CVPR 2023 • Wenjing Bian, ZiRui Wang, Kejie Li, Jia-Wang Bian, Victor Adrian Prisacariu
Recent advances in this direction demonstrate the possibility of jointly optimising a NeRF and camera poses in forward-facing scenes.
1 code implementation • 20 Jul 2022 • Qianyi Wu, Xian Liu, Yuedong Chen, Kejie Li, Chuanxia Zheng, Jianfei Cai, Jianmin Zheng
This paper proposes a novel framework, ObjectSDF, to build an object-compositional neural implicit representation with high fidelity in 3D reconstruction and object representation.
1 code implementation • CVPR 2022 • Kejie Li, Yansong Tang, Victor Adrian Prisacariu, Philip H. S. Torr
Dense 3D reconstruction from a stream of depth images is the key to many mixed reality and robotic applications.
1 code implementation • ICCV 2021 • Kejie Li, Daniel DeTone, Steven Chen, Minh Vo, Ian Reid, Hamid Rezatofighi, Chris Sweeney, Julian Straub, Richard Newcombe
Localizing objects and estimating their extent in 3D is an important step towards high-level 3D scene understanding, which has many applications in Augmented Reality and Robotics.
1 code implementation • 5 Jul 2021 • Wenjing Bian, ZiRui Wang, Kejie Li, Victor Adrian Prisacariu
We propose Ray-ONet to reconstruct detailed 3D models from monocular images efficiently.
no code implementations • 9 Dec 2020 • Kejie Li, Hamid Rezatofighi, Ian Reid
Given a new RGB frame, MOLTR firstly applies a monocular 3D detector to localise objects of interest and extract their shape codes that represent the object shapes in a learned embedding space.
no code implementations • 11 May 2020 • Kejie Li, Martin Rünz, Meng Tang, Lingni Ma, Chen Kong, Tanner Schmidt, Ian Reid, Lourdes Agapito, Julian Straub, Steven Lovegrove, Richard Newcombe
We introduce FroDO, a method for accurate 3D reconstruction of object instances from RGB video that infers object location, pose and shape in a coarse-to-fine manner.
no code implementations • 29 Nov 2018 • Kejie Li, Ravi Garg, Ming Cai, Ian Reid
3D shape reconstruction from a single image is a highly ill-posed problem.
no code implementations • 24 Sep 2018 • Mehdi Hosseinzadeh, Kejie Li, Yasir Latif, Ian Reid
While sparse point-based SLAM methods provide accurate camera localization, the generated maps lack semantic information.
no code implementations • ECCV 2018 • Kejie Li, Trung Pham, Huangying Zhan, Ian Reid
Given a single image at an arbitrary viewpoint, a CNN predicts multiple surfaces, each in a canonical location relative to the object.
1 code implementation • CVPR 2018 • Huangying Zhan, Ravi Garg, Chamara Saroj Weerasekera, Kejie Li, Harsh Agarwal, Ian Reid
Despite learning based methods showing promising results in single view depth estimation and visual odometry, most existing approaches treat the tasks in a supervised manner.