Search Results for author: Yufan Zhou

Found 20 papers, 8 papers with code

Towards Aligned Layout Generation via Diffusion Model with Aesthetic Constraints

1 code implementation7 Feb 2024 Jian Chen, Ruiyi Zhang, Yufan Zhou, Changyou Chen

Controllable layout generation refers to the process of creating a plausible visual arrangement of elements within a graphic design (e. g., document and web designs) with constraints representing design intentions.

Layout Design

Customization Assistant for Text-to-image Generation

1 code implementation5 Dec 2023 Yufan Zhou, Ruiyi Zhang, Jiuxiang Gu, Tong Sun

Some existing methods do not require fine-tuning, while their performance are unsatisfactory.

Descriptive Language Modelling +2

LLaVAR: Enhanced Visual Instruction Tuning for Text-Rich Image Understanding

1 code implementation29 Jun 2023 Yanzhe Zhang, Ruiyi Zhang, Jiuxiang Gu, Yufan Zhou, Nedim Lipka, Diyi Yang, Tong Sun

Instruction tuning unlocks the superior capability of Large Language Models (LLM) to interact with humans.

16k Image Captioning +3

Enhancing Detail Preservation for Customized Text-to-Image Generation: A Regularization-Free Approach

1 code implementation23 May 2023 Yufan Zhou, Ruiyi Zhang, Tong Sun, Jinhui Xu

However, generating images of novel concept provided by the user input image is still a challenging task.

Text-to-Image Generation

Towards Building the Federated GPT: Federated Instruction Tuning

1 code implementation9 May 2023 Jianyi Zhang, Saeed Vahidian, Martin Kuo, Chunyuan Li, Ruiyi Zhang, Tong Yu, Yufan Zhou, Guoyin Wang, Yiran Chen

This repository offers a foundational framework for exploring federated fine-tuning of LLMs using heterogeneous instructions across diverse categories.

Federated Learning

Shifted Diffusion for Text-to-image Generation

1 code implementation CVPR 2023 Yufan Zhou, Bingchen Liu, Yizhe Zhu, Xiao Yang, Changyou Chen, Jinhui Xu

Unlike the baseline diffusion model used in DALL-E 2, our method seamlessly encodes prior knowledge of the pre-trained CLIP model in its diffusion process by designing a new initialization distribution and a new transition step of the diffusion.

Zero-Shot Text-to-Image Generation

Lafite2: Few-shot Text-to-Image Generation

no code implementations25 Oct 2022 Yufan Zhou, Chunyuan Li, Changyou Chen, Jianfeng Gao, Jinhui Xu

The low requirement of the proposed method yields high flexibility and usability: it can be beneficial to a wide range of settings, including the few-shot, semi-supervised and fully-supervised learning; it can be applied on different models including generative adversarial networks (GANs) and diffusion models.

Retrieval Text-to-Image Generation

Towards Language-Free Training for Text-to-Image Generation

no code implementations CVPR 2022 Yufan Zhou, Ruiyi Zhang, Changyou Chen, Chunyuan Li, Chris Tensmeyer, Tong Yu, Jiuxiang Gu, Jinhui Xu, Tong Sun

One of the major challenges in training text-to-image generation models is the need of a large number of high-quality text-image pairs.

Zero-Shot Text-to-Image Generation

A Generic Approach for Enhancing GANs by Regularized Latent Optimization

no code implementations7 Dec 2021 Yufan Zhou, Chunyuan Li, Changyou Chen, Jinhui Xu

With the rapidly growing model complexity and data volume, training deep generative models (DGMs) for better performance has becoming an increasingly more important challenge.

Image Inpainting text-guided-image-editing +1

Learning High-Dimensional Distributions with Latent Neural Fokker-Planck Kernels

no code implementations10 May 2021 Yufan Zhou, Changyou Chen, Jinhui Xu

Learning high-dimensional distributions is an important yet challenging problem in machine learning with applications in various domains.

Vocal Bursts Intensity Prediction

Meta-Learning with Neural Tangent Kernels

no code implementations7 Feb 2021 Yufan Zhou, Zhenyi Wang, Jiayi Xian, Changyou Chen, Jinhui Xu

We achieve this goal by 1) replacing the adaptation with a fast-adaptive regularizer in the RKHS; and 2) solving the adaptation analytically based on the NTK theory.

Meta-Learning

Meta-Learning in Reproducing Kernel Hilbert Space

no code implementations ICLR 2021 Yufan Zhou, Zhenyi Wang, Jiayi Xian, Changyou Chen, Jinhui Xu

Within this paradigm, we introduce two meta learning algorithms in RKHS, which no longer need an explicit inner-loop adaptation as in the MAML framework.

Meta-Learning

Graph Neural Networks with Composite Kernels

no code implementations16 May 2020 Yufan Zhou, Jiayi Xian, Changyou Chen, Jinhui Xu

We then propose feature aggregation as the composition of the original neighbor-based kernel and a learnable kernel to encode feature similarities in a feature space.

Graph Attention

Learning Diverse Stochastic Human-Action Generators by Learning Smooth Latent Transitions

1 code implementation AAAI 2019 Zhenyi Wang, Ping Yu, Yang Zhao, Ruiyi Zhang, Yufan Zhou, Junsong Yuan, Changyou Chen

In this paper, we focus on skeleton-based action generation and propose to model smooth and diverse transitions on a latent space of action sequences with much lower dimensionality.

Action Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.