Search Results for author: Yongqi Yang

Found 4 papers, 2 papers with code

SOWing Information: Cultivating Contextual Coherence with MLLMs in Image Generation

no code implementations28 Nov 2024 Yuhan Pei, Ruoyu Wang, Yongqi Yang, Ye Zhu, Olga Russakovsky, Yu Wu

Originating from the diffusion phenomenon in physics, which describes the random movement and collisions of particles, diffusion generative models simulate a random walk in the data space along the denoising trajectory.

Denoising Image Generation

CoMM: A Coherent Interleaved Image-Text Dataset for Multimodal Understanding and Generation

1 code implementation15 Jun 2024 Wei Chen, Lin Li, Yongqi Yang, Bin Wen, Fan Yang, Tingting Gao, Yu Wu, Long Chen

To address this gap, we introduce CoMM, a high-quality Coherent interleaved image-text MultiModal dataset designed to enhance the coherence, consistency, and alignment of generated multimodal content.

In-Context Learning Visual Storytelling

Diffusion in Diffusion: Cyclic One-Way Diffusion for Text-Vision-Conditioned Generation

1 code implementation14 Jun 2023 Ruoyu Wang, Yongqi Yang, Zhihao Qian, Ye Zhu, Yu Wu

In this work, we investigate the diffusion (physics) in diffusion (machine learning) properties and propose our Cyclic One-Way Diffusion (COW) method to control the direction of diffusion phenomenon given a pre-trained frozen diffusion model for versatile customization application scenarios, where the low-level pixel information from the conditioning needs to be preserved.

Denoising Image Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.