Search Results for author: Ziyang Yuan

Found 9 papers, 5 papers with code

SmartEdit: Exploring Complex Instruction-based Image Editing with Multimodal Large Language Models

1 code implementation11 Dec 2023 Yuzhou Huang, Liangbin Xie, Xintao Wang, Ziyang Yuan, Xiaodong Cun, Yixiao Ge, Jiantao Zhou, Chao Dong, Rui Huang, Ruimao Zhang, Ying Shan

Both quantitative and qualitative results on this evaluation dataset indicate that our SmartEdit surpasses previous methods, paving the way for the practical application of complex instruction-based image editing.

MotionCtrl: A Unified and Flexible Motion Controller for Video Generation

1 code implementation6 Dec 2023 Zhouxia Wang, Ziyang Yuan, Xintao Wang, Tianshui Chen, Menghan Xia, Ping Luo, Ying Shan

Therefore, this paper presents MotionCtrl, a unified and flexible motion controller for video generation designed to effectively and independently control camera and object motion.

Object Video Generation

Phase Retrieval with Background Information: Decreased References and Efficient Methods

no code implementations16 Aug 2023 Ziyang Yuan, Haoxing Yang, Ningyi Leng, Hongxia Wang

Furthermore, two methods called Background Douglas-Rachford (BDR) and Convex Background Douglas-Rachford (CBDR) are proposed.


Untrained neural network embedded Fourier phase retrieval from few measurements

1 code implementation16 Jul 2023 Liyuan Ma, Hongxia Wang, Ningyi Leng, Ziyang Yuan

FPR with few measurements is important for reducing time and hardware costs, but it suffers from serious ill-posedness.


Make Encoder Great Again in 3D GAN Inversion through Geometry and Occlusion-Aware Encoding

no code implementations ICCV 2023 Ziyang Yuan, Yiming Zhu, Yu Li, Hongyu Liu, Chun Yuan

We leverage the inherent properties of EG3D's latent space to design a discriminator and a background depth regularization.

ADMM based Fourier phase retrieval with untrained generative prior

no code implementations23 Oct 2022 Liyuan Ma, Hongxia Wang, Ningyi Leng, Ziyang Yuan

Then an untrained generative network is embedded in the iterative process of ADMM to project an estimated signal into the generative space, and the projected signal is applied to next iteration of ADMM.


One Model to Edit Them All: Free-Form Text-Driven Image Manipulation with Semantic Modulations

1 code implementation14 Oct 2022 Yiming Zhu, Hongyu Liu, Yibing Song, Ziyang Yuan, Xintong Han, Chun Yuan, Qifeng Chen, Jue Wang

Based on the visual latent space of StyleGAN[21] and text embedding space of CLIP[34], studies focus on how to map these two latent spaces for text-driven attribute manipulations.

Attribute Image Manipulation

Cannot find the paper you are looking for? You can Submit a new open access paper.