Search Results for author: Miaomiao Cui

Found 28 papers, 16 papers with code

GaussianIP: Identity-Preserving Realistic 3D Human Generation via Human-Centric Diffusion Prior

1 code implementation14 Mar 2025 Zichen Tang, Yuan YAO, Miaomiao Cui, Liefeng Bo, Hongyu Yang

In stage 1, we propose a novel Adaptive Human Distillation Sampling (AHDS) method to rapidly generate a 3D human that maintains high identity consistency with the image prompt and achieves a realistic appearance.

Towards Fine-grained Interactive Segmentation in Images and Videos

no code implementations12 Feb 2025 Yuan YAO, Qiushi Yang, Miaomiao Cui, Liefeng Bo

Existing methods for high-precision interactive segmentation face a trade-off between the ability to perceive intricate local details and maintaining stable prompting capability, which hinders the applicability and effectiveness of foundational segmentation models.

Interactive Segmentation Segmentation

Perception-as-Control: Fine-grained Controllable Image Animation with 3D-aware Motion Representation

no code implementations9 Jan 2025 Yingjie Chen, Yifang Men, Yuan YAO, Miaomiao Cui, Liefeng Bo

Recent works have made progress in controlling camera or object motion via various motion representations, while they still struggle to support collaborative camera and object motion control with adaptive control granularity.

Image Animation Object

MIMO: Controllable Character Video Synthesis with Spatial Decomposed Modeling

no code implementations24 Sep 2024 Yifang Men, Yuan YAO, Miaomiao Cui, Liefeng Bo

To this end, we propose MIMO, a novel framework which can not only synthesize character videos with controllable attributes (i. e., character, motion and scene) provided by simple user inputs, but also simultaneously achieve advanced scalability to arbitrary characters, generality to novel 3D motions, and applicability to interactive real-world scenes in a unified framework.

MotionCom: Automatic and Motion-Aware Image Composition with LLM and Video Diffusion Prior

1 code implementation16 Sep 2024 Weijing Tao, Xiaofeng Yang, Miaomiao Cui, Guosheng Lin

This work presents MotionCom, a training-free motion-aware diffusion based image composition, enabling automatic and seamless integration of target objects into new scenes with dynamically coherent results without finetuning or optimization.

Image Generation Language Modeling +1

I4VGen: Image as Free Stepping Stone for Text-to-Video Generation

no code implementations4 Jun 2024 Xiefan Guo, Jinlin Liu, Miaomiao Cui, Liefeng Bo, Di Huang

Text-to-video generation has trailed behind text-to-image generation in terms of quality and diversity, primarily due to the inherent complexities of spatio-temporal modeling and the limited availability of video-text datasets.

Diversity Text-to-Image Generation +2

Disentangling Foreground and Background Motion for Enhanced Realism in Human Video Generation

no code implementations26 May 2024 Jinlin Liu, Kai Yu, Mengyang Feng, Xiefan Guo, Miaomiao Cui

Training on real-world videos enhanced with this innovative motion depiction approach, our model generates videos exhibiting coherent movement in both foreground subjects and their surrounding contexts.

Video Generation

InitNO: Boosting Text-to-Image Diffusion Models via Initial Noise Optimization

1 code implementation CVPR 2024 Xiefan Guo, Jinlin Liu, Miaomiao Cui, Jiankai Li, Hongyu Yang, Di Huang

Recent strides in the development of diffusion models, exemplified by advancements such as Stable Diffusion, have underscored their remarkable prowess in generating visually compelling images.

valid

DivAvatar: Diverse 3D Avatar Generation with a Single Prompt

no code implementations27 Feb 2024 Weijing Tao, Biwen Lei, Kunhao Liu, Shijian Lu, Miaomiao Cui, Xuansong Xie, Chunyan Miao

We design DivAvatar, a novel framework that generates diverse avatars, empowering 3D creatives with a multitude of distinct and richly varied 3D avatars from a single text prompt.

Diversity NeRF

3DToonify: Creating Your High-Fidelity 3D Stylized Avatar Easily from 2D Portrait Images

no code implementations CVPR 2024 Yifang Men, Hanxi Liu, Yuan YAO, Miaomiao Cui, Xuansong Xie, Zhouhui Lian

In this paper we make a connection between the two and tackle the challenging task of 3D portrait stylization - modeling high-fidelity 3D stylized avatars from captured 2D portrait images.

Style Transfer

DiffusionGAN3D: Boosting Text-guided 3D Generation and Domain Adaptation by Combining 3D GANs and Diffusion Priors

no code implementations CVPR 2024 Biwen Lei, Kai Yu, Mengyang Feng, Miaomiao Cui, Xuansong Xie

Extensive experiments demonstrate that the proposed framework achieves excellent results in both domain adaptation and text-to-avatar tasks, outperforming existing methods in terms of generation quality and efficiency.

3D Generation Domain Adaptation

DreaMoving: A Human Video Generation Framework based on Diffusion Models

no code implementations8 Dec 2023 Mengyang Feng, Jinlin Liu, Kai Yu, Yuan YAO, Zheng Hui, Xiefan Guo, Xianhui Lin, Haolan Xue, Chen Shi, Xiaowen Li, Aojie Li, Xiaoyang Kang, Biwen Lei, Miaomiao Cui, Peiran Ren, Xuansong Xie

In this paper, we present DreaMoving, a diffusion-based controllable video generation framework to produce high-quality customized human videos.

Video Generation

Boosting3D: High-Fidelity Image-to-3D by Boosting 2D Diffusion Prior to 3D Prior with Progressive Learning

no code implementations22 Nov 2023 Kai Yu, Jinlin Liu, Mengyang Feng, Miaomiao Cui, Xuansong Xie

After the progressive training, the LoRA learns the 3D information of the generated object and eventually turns to an object-level 3D prior.

3D Generation Image to 3D +2

Diffusion360: Seamless 360 Degree Panoramic Image Generation based on Diffusion Models

1 code implementation22 Nov 2023 Mengyang Feng, Jinlin Liu, Miaomiao Cui, Xuansong Xie

This is a technical report on the 360-degree panoramic image generation task based on diffusion models.

Denoising Image Generation

Box2Mask: Box-supervised Instance Segmentation via Level-set Evolution

2 code implementations3 Dec 2022 Wentong Li, Wenyu Liu, Jianke Zhu, Miaomiao Cui, Risheng Yu, Xiansheng Hua, Lei Zhang

In contrast to fully supervised methods using pixel-wise mask labels, box-supervised instance segmentation takes advantage of simple box annotations, which has recently attracted increasing research attention.

Box-supervised Instance Segmentation Decoder +1

Box-supervised Instance Segmentation with Level Set Evolution

2 code implementations19 Jul 2022 Wentong Li, Wenyu Liu, Jianke Zhu, Miaomiao Cui, Xiansheng Hua, Lei Zhang

A simple mask supervised SOLOv2 model is adapted to predict the instance-aware mask map as the level set for each instance.

Box-supervised Instance Segmentation Segmentation

DCT-Net: Domain-Calibrated Translation for Portrait Stylization

3 code implementations6 Jul 2022 Yifang Men, Yuan YAO, Miaomiao Cui, Zhouhui Lian, Xuansong Xie

This paper introduces DCT-Net, a novel image translation architecture for few-shot portrait stylization.

Few-Shot Learning Style Transfer +1

Towards Counterfactual Image Manipulation via CLIP

1 code implementation6 Jul 2022 Yingchen Yu, Fangneng Zhan, Rongliang Wu, Jiahui Zhang, Shijian Lu, Miaomiao Cui, Xuansong Xie, Xian-Sheng Hua, Chunyan Miao

In addition, we design a simple yet effective scheme that explicitly maps CLIP embeddings (of target text) to the latent space and fuses them with latent codes for effective latent code optimization and accurate editing.

counterfactual Image Manipulation

Improving Nighttime Driving-Scene Segmentation via Dual Image-adaptive Learnable Filters

2 code implementations4 Jul 2022 Wenyu Liu, Wentong Li, Jianke Zhu, Miaomiao Cui, Xuansong Xie, Lei Zhang

With DIAL-Filters, we design both unsupervised and supervised frameworks for nighttime driving-scene segmentation, which can be trained in an end-to-end manner.

Autonomous Driving Scene Segmentation +1

ABPN: Adaptive Blend Pyramid Network for Real-Time Local Retouching of Ultra High-Resolution Photo

1 code implementation CVPR 2022 Biwen Lei, Xiefan Guo, Hongyu Yang, Miaomiao Cui, Xuansong Xie, Di Huang

The network is mainly composed of two components: a context-aware local retouching layer (LRL) and an adaptive blend pyramid layer (BPL).

4k Photo Retouching

Unpaired Cartoon Image Synthesis via Gated Cycle Mapping

no code implementations CVPR 2022 Yifang Men, Yuan YAO, Miaomiao Cui, Zhouhui Lian, Xuansong Xie, Xian-Sheng Hua

Experimental results demonstrate the superiority of the proposed method over the state of the art and validate its effectiveness in the brand-new task of general cartoon image synthesis.

Image Generation Video Generation

Attention-guided Temporally Coherent Video Object Matting

1 code implementation24 May 2021 Yunke Zhang, Chi Wang, Miaomiao Cui, Peiran Ren, Xuansong Xie, Xian-Sheng Hua, Hujun Bao, QiXing Huang, Weiwei Xu

Experimental results show that our method can generate high-quality alpha mattes for various videos featuring appearance change, occlusion, and fast motion.

Image Matting Object +4

PPR10K: A Large-Scale Portrait Photo Retouching Dataset with Human-Region Mask and Group-Level Consistency

1 code implementation CVPR 2021 Jie Liang, Hui Zeng, Miaomiao Cui, Xuansong Xie, Lei Zhang

HRP requires that more attention should be paid to human regions, while GLC requires that a group of portrait photos should be retouched to a consistent tone.

Photo Retouching

Boosting Semantic Human Matting with Coarse Annotations

1 code implementation CVPR 2020 Jinlin Liu, Yuan YAO, Wendi Hou, Miaomiao Cui, Xuansong Xie, Chang-Shui Zhang, Xian-Sheng Hua

In this paper, we propose to use coarse annotated data coupled with fine annotated data to boost end-to-end semantic human matting without trimaps as extra input.

Image Matting Semantic Segmentation

Cannot find the paper you are looking for? You can Submit a new open access paper.