no code implementations • 25 Mar 2024 • Jingyuan Zhu, Huimin Ma, Jiansheng Chen, Jian Yuan
This paper presents a general approach for text-to-image diffusion models to address the mutual interference between different subjects and their attachments in complex scenes, pursuing better text-image consistency.
no code implementations • 25 Jun 2023 • Jingyuan Zhu, Huimin Ma, Jiansheng Chen, Jian Yuan
Typical diffusion models and modern large-scale conditional generative models like text-to-image generative models are vulnerable to overfitting when fine-tuned on extremely limited data.
no code implementations • 19 May 2023 • Jingyuan Zhu, Huimin Ma, Jiansheng Chen, Jian Yuan
Our approach only needs the silhouettes of few-shot target samples as training data to learn target geometry distributions and achieve generated shapes with diverse topology and textures.
1 code implementation • 6 Mar 2023 • Jingyuan Zhu, Huimin Ma, Jiansheng Chen, Jian Yuan
We present MotionVideoGAN, a novel video generator synthesizing videos based on the motion space learned by pre-trained image pair generators.
no code implementations • 7 Nov 2022 • Jingyuan Zhu, Huimin Ma, Jiansheng Chen, Jian Yuan
Then we fine-tune DDPMs pre-trained on large source domains to solve the overfitting problem when training data is limited.
no code implementations • 27 Oct 2022 • Jingyuan Zhu, Huimin Ma, Jiansheng Chen, Jian Yuan
It strengthens global image discrimination and guides adapted GANs to preserve more information learned from source domains for higher image quality.