1 code implementation • 14 Mar 2025 • Zichen Tang, Yuan YAO, Miaomiao Cui, Liefeng Bo, Hongyu Yang
In stage 1, we propose a novel Adaptive Human Distillation Sampling (AHDS) method to rapidly generate a 3D human that maintains high identity consistency with the image prompt and achieves a realistic appearance.
no code implementations • 12 Feb 2025 • Yuan YAO, Qiushi Yang, Miaomiao Cui, Liefeng Bo
Existing methods for high-precision interactive segmentation face a trade-off between the ability to perceive intricate local details and maintaining stable prompting capability, which hinders the applicability and effectiveness of foundational segmentation models.
no code implementations • 9 Jan 2025 • Yingjie Chen, Yifang Men, Yuan YAO, Miaomiao Cui, Liefeng Bo
Recent works have made progress in controlling camera or object motion via various motion representations, while they still struggle to support collaborative camera and object motion control with adaptive control granularity.
no code implementations • 24 Sep 2024 • Yifang Men, Yuan YAO, Miaomiao Cui, Liefeng Bo
To this end, we propose MIMO, a novel framework which can not only synthesize character videos with controllable attributes (i. e., character, motion and scene) provided by simple user inputs, but also simultaneously achieve advanced scalability to arbitrary characters, generality to novel 3D motions, and applicability to interactive real-world scenes in a unified framework.
1 code implementation • 16 Sep 2024 • Weijing Tao, Xiaofeng Yang, Miaomiao Cui, Guosheng Lin
This work presents MotionCom, a training-free motion-aware diffusion based image composition, enabling automatic and seamless integration of target objects into new scenes with dynamically coherent results without finetuning or optimization.
no code implementations • 4 Jun 2024 • Xiefan Guo, Jinlin Liu, Miaomiao Cui, Liefeng Bo, Di Huang
Text-to-video generation has trailed behind text-to-image generation in terms of quality and diversity, primarily due to the inherent complexities of spatio-temporal modeling and the limited availability of video-text datasets.
no code implementations • 26 May 2024 • Jinlin Liu, Kai Yu, Mengyang Feng, Xiefan Guo, Miaomiao Cui
Training on real-world videos enhanced with this innovative motion depiction approach, our model generates videos exhibiting coherent movement in both foreground subjects and their surrounding contexts.
1 code implementation • CVPR 2024 • Xiefan Guo, Jinlin Liu, Miaomiao Cui, Jiankai Li, Hongyu Yang, Di Huang
Recent strides in the development of diffusion models, exemplified by advancements such as Stable Diffusion, have underscored their remarkable prowess in generating visually compelling images.
no code implementations • 27 Feb 2024 • Weijing Tao, Biwen Lei, Kunhao Liu, Shijian Lu, Miaomiao Cui, Xuansong Xie, Chunyan Miao
We design DivAvatar, a novel framework that generates diverse avatars, empowering 3D creatives with a multitude of distinct and richly varied 3D avatars from a single text prompt.
no code implementations • CVPR 2024 • Yifang Men, Biwen Lei, Yuan YAO, Miaomiao Cui, Zhouhui Lian, Xuansong Xie
We present En3D, an enhanced generative scheme for sculpting high-quality 3D human avatars.
no code implementations • CVPR 2024 • Yifang Men, Hanxi Liu, Yuan YAO, Miaomiao Cui, Xuansong Xie, Zhouhui Lian
In this paper we make a connection between the two and tackle the challenging task of 3D portrait stylization - modeling high-fidelity 3D stylized avatars from captured 2D portrait images.
no code implementations • CVPR 2024 • Biwen Lei, Kai Yu, Mengyang Feng, Miaomiao Cui, Xuansong Xie
Extensive experiments demonstrate that the proposed framework achieves excellent results in both domain adaptation and text-to-avatar tasks, outperforming existing methods in terms of generation quality and efficiency.
no code implementations • 8 Dec 2023 • Mengyang Feng, Jinlin Liu, Kai Yu, Yuan YAO, Zheng Hui, Xiefan Guo, Xianhui Lin, Haolan Xue, Chen Shi, Xiaowen Li, Aojie Li, Xiaoyang Kang, Biwen Lei, Miaomiao Cui, Peiran Ren, Xuansong Xie
In this paper, we present DreaMoving, a diffusion-based controllable video generation framework to produce high-quality customized human videos.
no code implementations • 22 Nov 2023 • Kai Yu, Jinlin Liu, Mengyang Feng, Miaomiao Cui, Xuansong Xie
After the progressive training, the LoRA learns the 3D information of the generated object and eventually turns to an object-level 3D prior.
1 code implementation • 22 Nov 2023 • Mengyang Feng, Jinlin Liu, Miaomiao Cui, Xuansong Xie
This is a technical report on the 360-degree panoramic image generation task based on diffusion models.
1 code implementation • CVPR 2023 • Biwen Lei, Jianqiang Ren, Mengyang Feng, Miaomiao Cui, Xuansong Xie
Meanwhile, 3D priors of facial details are incorporated to enhance the accuracy and authenticity of the reconstruction results.
Ranked #4 on
3D Face Reconstruction
on REALY (side-view)
2 code implementations • 3 Dec 2022 • Wentong Li, Wenyu Liu, Jianke Zhu, Miaomiao Cui, Risheng Yu, Xiansheng Hua, Lei Zhang
In contrast to fully supervised methods using pixel-wise mask labels, box-supervised instance segmentation takes advantage of simple box annotations, which has recently attracted increasing research attention.
2 code implementations • 19 Jul 2022 • Wentong Li, Wenyu Liu, Jianke Zhu, Miaomiao Cui, Xiansheng Hua, Lei Zhang
A simple mask supervised SOLOv2 model is adapted to predict the instance-aware mask map as the level set for each instance.
3 code implementations • 6 Jul 2022 • Yifang Men, Yuan YAO, Miaomiao Cui, Zhouhui Lian, Xuansong Xie
This paper introduces DCT-Net, a novel image translation architecture for few-shot portrait stylization.
1 code implementation • 6 Jul 2022 • Yingchen Yu, Fangneng Zhan, Rongliang Wu, Jiahui Zhang, Shijian Lu, Miaomiao Cui, Xuansong Xie, Xian-Sheng Hua, Chunyan Miao
In addition, we design a simple yet effective scheme that explicitly maps CLIP embeddings (of target text) to the latent space and fuses them with latent codes for effective latent code optimization and accurate editing.
2 code implementations • 4 Jul 2022 • Wenyu Liu, Wentong Li, Jianke Zhu, Miaomiao Cui, Xuansong Xie, Lei Zhang
With DIAL-Filters, we design both unsupervised and supervised frameworks for nighttime driving-scene segmentation, which can be trained in an end-to-end manner.
1 code implementation • CVPR 2022 • Jianqiang Ren, Yuan YAO, Biwen Lei, Miaomiao Cui, Xuansong Xie
Body reshaping is an important procedure in portrait photo retouching.
1 code implementation • CVPR 2022 • Biwen Lei, Xiefan Guo, Hongyu Yang, Miaomiao Cui, Xuansong Xie, Di Huang
The network is mainly composed of two components: a context-aware local retouching layer (LRL) and an adaptive blend pyramid layer (BPL).
no code implementations • CVPR 2022 • Yifang Men, Yuan YAO, Miaomiao Cui, Zhouhui Lian, Xuansong Xie, Xian-Sheng Hua
Experimental results demonstrate the superiority of the proposed method over the state of the art and validate its effectiveness in the brand-new task of general cartoon image synthesis.
1 code implementation • 24 May 2021 • Yunke Zhang, Chi Wang, Miaomiao Cui, Peiran Ren, Xuansong Xie, Xian-Sheng Hua, Hujun Bao, QiXing Huang, Weiwei Xu
Experimental results show that our method can generate high-quality alpha mattes for various videos featuring appearance change, occlusion, and fast motion.
1 code implementation • CVPR 2021 • Jie Liang, Hui Zeng, Miaomiao Cui, Xuansong Xie, Lei Zhang
HRP requires that more attention should be paid to human regions, while GLC requires that a group of portrait photos should be retouched to a consistent tone.
1 code implementation • 4 Feb 2021 • Chi Wang, Yunke Zhang, Miaomiao Cui, Peiran Ren, Yin Yang, Xuansong Xie, Xiansheng Hua, Hujun Bao, Weiwei Xu
This paper proposes a novel active boundary loss for semantic segmentation.
1 code implementation • CVPR 2020 • Jinlin Liu, Yuan YAO, Wendi Hou, Miaomiao Cui, Xuansong Xie, Chang-Shui Zhang, Xian-Sheng Hua
In this paper, we propose to use coarse annotated data coupled with fine annotated data to boost end-to-end semantic human matting without trimaps as extra input.
Ranked #9 on
Image Matting
on AM-2K