Search Results for author: Xuanchi Ren

Found 10 papers, 9 papers with code

XCube ($\mathcal{X}^3$): Large-Scale 3D Generative Modeling using Sparse Voxel Hierarchies

no code implementations6 Dec 2023 Xuanchi Ren, Jiahui Huang, Xiaohui Zeng, Ken Museth, Sanja Fidler, Francis Williams

In addition to unconditional generation, we show that our model can be used to solve a variety of tasks such as user-guided editing, scene completion from a single scan, and text-to-3D.

3D Shape Generation Scene Generation +1

Blind Video Deflickering by Neural Filtering with a Flawed Atlas

1 code implementation CVPR 2023 Chenyang Lei, Xuanchi Ren, Zhaoxiang Zhang, Qifeng Chen

Prior work usually requires specific guidance such as the flickering frequency, manual annotations, or extra consistent videos to remove the flicker.

Video Generation Video Temporal Consistency

Look Outside the Room: Synthesizing A Consistent Long-Term 3D Scene Video from A Single Image

1 code implementation CVPR 2022 Xuanchi Ren, Xiaolong Wang

Novel view synthesis from a single image has recently attracted a lot of attention, and it has been primarily advanced by 3D deep learning and rendering techniques.

Novel View Synthesis

Retriever: Learning Content-Style Representation as a Token-Level Bipartite Graph

2 code implementations ICLR 2022 Dacheng Yin, Xuanchi Ren, Chong Luo, Yuwang Wang, Zhiwei Xiong, Wenjun Zeng

Last, an innovative link attention module serves as the decoder to reconstruct data from the decomposed content and style, with the help of the linking keys.

Quantization Style Transfer +1

Rethinking Content and Style: Exploring Bias for Unsupervised Disentanglement

1 code implementation21 Feb 2021 Xuanchi Ren, Tao Yang, Yuwang Wang, Wenjun Zeng

From the unsupervised disentanglement perspective, we rethink content and style and propose a formulation for unsupervised C-S disentanglement based on our assumption that different factors are of different importance and popularity for image reconstruction, which serves as a data bias.

3D Reconstruction Disentanglement +4

Learning Disentangled Representation by Exploiting Pretrained Generative Models: A Contrastive Learning View

2 code implementations ICLR 2022 Xuanchi Ren, Tao Yang, Yuwang Wang, Wenjun Zeng

Based on this observation, we argue that it is possible to mitigate the trade-off by $(i)$ leveraging the pretrained generative models with high generation quality, $(ii)$ focusing on discovering the traversal directions as factors for disentangled representation learning.

Contrastive Learning Disentanglement

Towards Building A Group-based Unsupervised Representation Disentanglement Framework

1 code implementation ICLR 2022 Tao Yang, Xuanchi Ren, Yuwang Wang, Wenjun Zeng, Nanning Zheng

We then propose a model, based on existing VAE-based methods, to tackle the unsupervised learning problem of the framework.

Disentanglement

Video Deblurring by Fitting to Test Data

1 code implementation9 Dec 2020 Xuanchi Ren, Zian Qian, Qifeng Chen

Our key observation is that some frames in a video with motion blur are much sharper than others, and thus we can transfer the texture information in those sharp frames to blurry frames.

Autonomous Vehicles Deblurring

Music-oriented Dance Video Synthesis with Pose Perceptual Loss

1 code implementation13 Dec 2019 Xuanchi Ren, Haoran Li, Zijian Huang, Qifeng Chen

We present a learning-based approach with pose perceptual loss for automatic music video generation.

Video Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.