no code implementations • 26 Sep 2024 • Jiaxiang Tang, Zhaoshuo Li, Zekun Hao, Xian Liu, Gang Zeng, Ming-Yu Liu, Qinsheng Zhang
Current auto-regressive mesh generation methods suffer from issues such as incompleteness, insufficient detail, and poor generalization.
1 code implementation • 6 Jun 2024 • Diwen Wan, Ruijie Lu, Gang Zeng
Empowered by these superpoints, our method manages to extend 3D Gaussian splatting to dynamic scenes with only a slight increase in computational expense.
no code implementations • 4 Jun 2024 • Qi Wang, Ruijie Lu, Xudong Xu, Jingbo Wang, Michael Yu Wang, Bo Dai, Gang Zeng, Dan Xu
In the coarse stage, RoomTex first unwraps the scene mesh to a panoramic depth map and leverages ControlNet to generate a room panorama, which is regarded as the coarse reference to ensure the global texture consistency.
no code implementations • 18 Mar 2024 • Jiaxiang Tang, Ruijie Lu, Xiaokang Chen, Xiang Wen, Gang Zeng, Ziwei Liu
Text-to-texture synthesis has become a new frontier in 3D content creation thanks to the recent advances in text-to-image models.
1 code implementation • 7 Feb 2024 • Jiaxiang Tang, Zhaoxi Chen, Xiaokang Chen, Tengfei Wang, Gang Zeng, Ziwei Liu
2) 3D Backbone: We present an asymmetric U-Net as a high-throughput backbone operating on multi-view images, which can be produced from text or single-view image input by leveraging multi-view diffusion models.
1 code implementation • 28 Dec 2023 • Jiawei Ren, Liang Pan, Jiaxiang Tang, Chi Zhang, Ang Cao, Gang Zeng, Ziwei Liu
Specifically, we propose an integral framework with two major modules: 1) Image-to-4D GS - we initially generate static GS with DreamGaussianHD, followed by HexPlane-based dynamic generation with Gaussian deformation; and 2) Video-to-Video Texture Refinement - we refine the generated UV-space texture maps and meanwhile enhance their temporal consistency by utilizing a pre-trained image-to-video diffusion model.
no code implementations • CVPR 2024 • Xian Liu, Xiaohang Zhan, Jiaxiang Tang, Ying Shan, Gang Zeng, Dahua Lin, Xihui Liu, Ziwei Liu
In this paper, we propose an efficient yet effective framework, HumanGaussian, that generates high-quality 3D humans with fine-grained geometry and realistic appearance.
1 code implementation • 28 Sep 2023 • Jiaxiang Tang, Jiawei Ren, Hang Zhou, Ziwei Liu, Gang Zeng
In contrast to the occupancy pruning used in Neural Radiance Fields, we demonstrate that the progressive densification of 3D Gaussians converges significantly faster for 3D generative tasks.
no code implementations • 25 May 2023 • Xiaokang Chen, Jiaxiang Tang, Diwen Wan, Jingbo Wang, Gang Zeng
We propose to imitate the backbone feature of off-the-shelf perception models to achieve zero-shot semantic segmentation with NeRF.
2 code implementations • NeurIPS 2023 • Wenhai Wang, Zhe Chen, Xiaokang Chen, Jiannan Wu, Xizhou Zhu, Gang Zeng, Ping Luo, Tong Lu, Jie zhou, Yu Qiao, Jifeng Dai
We hope this model can set a new baseline for generalist vision and language models.
no code implementations • 20 Mar 2023 • Xiaokang Chen, Yajie Xing, Gang Zeng
In this paper, we propose a real-time semantic scene completion method with a feature aggregation strategy and conditioned prediction module.
1 code implementation • ICCV 2023 • Jiaxiang Tang, Hang Zhou, Xiaokang Chen, Tianshu Hu, Errui Ding, Jingdong Wang, Gang Zeng
Neural Radiance Fields (NeRF) have constituted a remarkable breakthrough in image-based 3D reconstruction.
1 code implementation • 8 Feb 2023 • Chao Chen, Haoyu Geng, Gang Zeng, Zhaobing Han, Hua Chai, Xiaokang Yang, Junchi Yan
Inductive one-bit matrix completion is motivated by modern applications such as recommender systems, where new users would appear at test stage with the ratings consisting of only ones and no zeros.
1 code implementation • 22 Nov 2022 • Jiaxiang Tang, Kaisiyuan Wang, Hang Zhou, Xiaokang Chen, Dongliang He, Tianshu Hu, Jingtuo Liu, Gang Zeng, Jingdong Wang
While dynamic Neural Radiance Fields (NeRF) have shown success in high-fidelity 3D modeling of talking portraits, the slow training and inference speed severely obstruct their potential usage.
no code implementations • 17 Nov 2022 • Xiaokang Chen, Jiahui Chen, Yan Liu, Gang Zeng
Specifically, Adaptive Matching applies bipartite matching to adaptively match the outputs of the teacher and the student in each decoder layer, while Fixed Matching fixes the correspondence between the outputs of the teacher and the student with the same object queries, with the teacher's fixed object queries fed to the decoder of the student as an auxiliary group.
no code implementations • 21 Aug 2022 • Longrui Dong, Gang Zeng
The ability for a moving agent to localize itself in environment is the basic demand for emerging applications, such as autonomous driving, etc.
2 code implementations • ICCV 2023 • Qiang Chen, Xiaokang Chen, Jian Wang, Shan Zhang, Kun Yao, Haocheng Feng, Junyu Han, Errui Ding, Gang Zeng, Jingdong Wang
Detection transformer (DETR) relies on one-to-one assignment, assigning one ground-truth object to one prediction, for end-to-end detection without NMS post-processing.
no code implementations • 18 Jul 2022 • Xiaokang Chen, Fangyun Wei, Gang Zeng, Jingdong Wang
Inspired by Conditional DETR, an improved DETR with fast training convergence, that presented box queries (originally called spatial queries) for internal decoder layers, we reformulate the object query into the format of the box query that is a composition of the embeddings of the reference point and the transformation of the box with respect to the reference point.
2 code implementations • 30 May 2022 • Jiaxiang Tang, Xiaokang Chen, Jingbo Wang, Gang Zeng
To circumvent the hurdle, in this paper, we present an explicit neural field representation that enables efficient and convenient manipulation of models.
1 code implementation • 31 Mar 2022 • Jiaxiang Tang, Xiaokang Chen, Jingbo Wang, Gang Zeng
Semantic scene reconstruction from point cloud is an essential and challenging task for 3D scene understanding.
no code implementations • 28 Mar 2022 • Min Zhong, Xinghao Chen, Xiaokang Chen, Gang Zeng, Yunhe Wang
For instance, our approach achieves a 66. 4\% mAP with the 0. 5 IoU threshold on the ScanNetV2 test set, which is 1. 9\% higher than the state-of-the-art method.
Ranked #7 on 3D Instance Segmentation on S3DIS
6 code implementations • 7 Feb 2022 • Xiaokang Chen, Mingyu Ding, Xiaodi Wang, Ying Xin, Shentong Mo, Yunhao Wang, Shumin Han, Ping Luo, Gang Zeng, Jingdong Wang
The pretraining tasks include two tasks: masked representation prediction - predict the representations for the masked patches, and masked patch reconstruction - reconstruct the masked patches.
no code implementations • 24 Dec 2021 • Xiaokang Chen, Jiaxiang Tang, Jingbo Wang, Gang Zeng
Firstly, we transfer the voxelized scenes to point clouds by removing these visible empty voxels and adopt a deep point stream to capture semantic information from the scene efficiently.
Ranked #5 on 3D Semantic Scene Completion on NYUv2
4 code implementations • ICCV 2021 • Depu Meng, Xiaokang Chen, Zejia Fan, Gang Zeng, Houqiang Li, Yuhui Yuan, Lei Sun, Jingdong Wang
Our approach, named conditional DETR, learns a conditional spatial query from the decoder embedding for decoder multi-head cross-attention.
Ranked #108 on Object Detection on COCO minival
1 code implementation • 19 Jul 2021 • Jiaxiang Tang, Xiaokang Chen, Gang Zeng
Inspired by the recent progress in implicit neural representation, we propose to formulate the guided super-resolution as a neural implicit image interpolation problem, where we take the form of a general image interpolation but use a novel Joint Implicit Image Function (JIIF) representation to learn both the interpolation weights and values.
3 code implementations • CVPR 2021 • Xiaokang Chen, Yuhui Yuan, Gang Zeng, Jingdong Wang
Our approach imposes the consistency on two segmentation networks perturbed with different initialization for the same input image.
Ranked #2 on Semi-Supervised Semantic Segmentation on WoodScape
no code implementations • ECAI 2020 • Min Zhong, Gang Zeng
In this work, a Semantic Point Completion Network (SPCNet) is proposed to address SSC in the point cloud space.
Ranked #9 on 3D Semantic Scene Completion on NYUv2
2 code implementations • ECCV 2020 • Yajie Xing, Jingbo Wang, Gang Zeng
In this paper, we propose a novel operator called malleable 2. 5D convolution to learn the receptive field along the depth-axis.
Ranked #52 on Semantic Segmentation on NYU Depth v2
2 code implementations • ECCV 2020 • Xiaokang Chen, Kwan-Yee Lin, Jingbo Wang, Wayne Wu, Chen Qian, Hongsheng Li, Gang Zeng
Depth information has proven to be a useful cue in the semantic segmentation of RGB-D images for providing a geometric counterpart to the RGB representation.
Ranked #1 on Semantic Segmentation on TLCGIS
2 code implementations • CVPR 2020 • Xiaokang Chen, Kwan-Yee Lin, Chen Qian, Gang Zeng, Hongsheng Li
To this end, we first propose a novel 3D sketch-aware feature embedding to explicitly encode geometric information effectively and efficiently.
3D Semantic Scene Completion from a single RGB image Hallucination
no code implementations • CVPR 2018 • Falong Shen, Shuicheng Yan, Gang Zeng
Recent works on style transfer typically need to train image transformation networks for every new style, and the style is encoded in the network parameters by enormous iterations of stochastic gradient descent, which lacks the generalization ability to new style in the inference stage.
1 code implementation • 13 Sep 2017 • Falong Shen, Shuicheng Yan, Gang Zeng
Recent works on style transfer typically need to train image transformation networks for every new style, and the style is encoded in the network parameters by enormous iterations of stochastic gradient descent.
1 code implementation • CVPR 2017 • Falong Shen, Rui Gan, Shuicheng Yan, Gang Zeng
The proposed joint model also employs a guidance CRF to further enhance the segmentation performance.
no code implementations • 28 May 2016 • Falong Shen, Gang Zeng
The weighted residual network is able to learn to combine residuals from different layers effectively and efficiently.
no code implementations • 13 May 2016 • Falong Shen, Gang Zeng
This paper describes a fast and accurate semantic image segmentation approach that encodes not only the discriminative features from deep neural networks, but also the high-order context compatibility among adjacent objects as well as low level image features.
no code implementations • CVPR 2014 • Jing Li, Zhichao Lu, Gang Zeng, Rui Gan, Hongbin Zha
This paper describes a patchwork assembly algorithm for depth image super-resolution.
no code implementations • 11 Dec 2013 • Jingdong Wang, Jing Wang, Qifa Ke, Gang Zeng, Shipeng Li
Traditional $k$-means is an iterative algorithm---in each iteration new cluster centers are computed and each data point is re-assigned to its nearest center.
no code implementations • 11 Dec 2013 • Jingdong Wang, Jing Wang, Gang Zeng, Rui Gan, Shipeng Li, Baining Guo
This structure augments the neighborhood graph with a bridge graph.
no code implementations • 30 Jul 2013 • Jingdong Wang, Jing Wang, Gang Zeng, Zhuowen Tu, Rui Gan, Shipeng Li
The $k$-NN graph has played a central role in increasingly popular data-driven techniques for various learning and vision tasks; yet, finding an efficient and effective way to construct $k$-NN graphs remains a challenge, especially for large-scale high-dimensional data.
no code implementations • CVPR 2013 • Peng Wang, Jingdong Wang, Gang Zeng, Weiwei Xu, Hongbin Zha, Shipeng Li
In visual recognition tasks, the design of low level image feature representation is fundamental.