1 code implementation • 11 Dec 2023 • Yuzhou Huang, Liangbin Xie, Xintao Wang, Ziyang Yuan, Xiaodong Cun, Yixiao Ge, Jiantao Zhou, Chao Dong, Rui Huang, Ruimao Zhang, Ying Shan
Both quantitative and qualitative results on this evaluation dataset indicate that our SmartEdit surpasses previous methods, paving the way for the practical application of complex instruction-based image editing.
1 code implementation • 6 Dec 2023 • Zhouxia Wang, Ziyang Yuan, Xintao Wang, Tianshui Chen, Menghan Xia, Ping Luo, Ying Shan
Therefore, this paper presents MotionCtrl, a unified and flexible motion controller for video generation designed to effectively and independently control camera and object motion.
no code implementations • 30 Oct 2023 • Ziyang Yuan, Mingdeng Cao, Xintao Wang, Zhongang Qi, Chun Yuan, Ying Shan
As a result, our CustomNet ensures enhanced identity preservation and generates diverse, harmonious outputs.
no code implementations • 16 Aug 2023 • Ziyang Yuan, Haoxing Yang, Ningyi Leng, Hongxia Wang
Furthermore, two methods called Background Douglas-Rachford (BDR) and Convex Background Douglas-Rachford (CBDR) are proposed.
1 code implementation • 16 Jul 2023 • Liyuan Ma, Hongxia Wang, Ningyi Leng, Ziyang Yuan
FPR with few measurements is important for reducing time and hardware costs, but it suffers from serious ill-posedness.
no code implementations • ICCV 2023 • Ziyang Yuan, Yiming Zhu, Yu Li, Hongyu Liu, Chun Yuan
We leverage the inherent properties of EG3D's latent space to design a discriminator and a background depth regularization.
no code implementations • 23 Oct 2022 • Liyuan Ma, Hongxia Wang, Ningyi Leng, Ziyang Yuan
Then an untrained generative network is embedded in the iterative process of ADMM to project an estimated signal into the generative space, and the projected signal is applied to next iteration of ADMM.
1 code implementation • 14 Oct 2022 • Yiming Zhu, Hongyu Liu, Yibing Song, Ziyang Yuan, Xintong Han, Chun Yuan, Qifeng Chen, Jue Wang
Based on the visual latent space of StyleGAN[21] and text embedding space of CLIP[34], studies focus on how to map these two latent spaces for text-driven attribute manipulations.
2 code implementations • 7 Mar 2020 • Jie Chen, Ziyang Yuan, Jian Peng, Li Chen, Haozhe Huang, Jiawei Zhu, Yu Liu, Haifeng Li
However, the available methods focus mainly on the difference information between multitemporal remote sensing images and lack robustness to pseudo-change information.