Layout-to-Image Generation

18 papers with code • 7 benchmarks • 4 datasets

Layout-to-image generation its the task to generate a scene based on the given layout. The layout describes the location of the objects to be included in the output image. In this section, you can find state-of-the-art leaderboards for Layout-to-image generation.

DivCon: Divide and Conquer for Progressive Text-to-Image Generation

divcon-gen/divcon 11 Mar 2024

To further improve T2I models' capability in numerical and spatial reasoning, the layout is employed as an intermedium to bridge large language models and layout-based diffusion models.

8
11 Mar 2024

Adversarial Supervision Makes Layout-to-Image Diffusion Models Thrive

boschresearch/aldm 16 Jan 2024

Current L2I models either suffer from poor editability via text or weak alignment between the generated image and the input layout.

40
16 Jan 2024

LLM Blueprint: Enabling Text-to-Image Generation with Complex and Detailed Prompts

hananshafi/llmblueprint 16 Oct 2023

Diffusion-based generative models have significantly advanced text-to-image generation but encounter challenges when processing lengthy and intricate text prompts describing complex scenes with multiple objects.

46
16 Oct 2023

Diagnostic Benchmark and Iterative Inpainting for Layout-Guided Image Generation

j-min/IterInpaint 13 Apr 2023

In this paper, we propose LayoutBench, a diagnostic benchmark for layout-guided image generation that examines four categories of spatial control skills: number, position, size, and shape.

21
13 Apr 2023

LayoutDiffusion: Controllable Diffusion Model for Layout-to-image Generation

zgctroy/layoutdiffusion CVPR 2023

To overcome the difficult multimodal fusion of image and layout, we propose to construct a structural image patch with region information and transform the patched image into a special layout to fuse with the normal layout in a unified form.

222
30 Mar 2023

Freestyle Layout-to-Image Synthesis

essunny310/freestylenet CVPR 2023

In this work, we explore the freestyle capability of the model, i. e., how far can it generate unseen semantics (e. g., classes, attributes, and styles) onto a given layout, and call the task Freestyle LIS (FLIS).

131
25 Mar 2023

Modeling Image Composition for Complex Scene Generation

johndreamer/twfa CVPR 2022

Compared to existing CNN-based and Transformer-based generation models that entangled modeling on pixel-level&patch-level and object-level&patch-level respectively, the proposed focal attention predicts the current patch token by only focusing on its highly-related tokens that specified by the spatial layout, thereby achieving disambiguation during training.

9
02 Jun 2022

Interactive Image Synthesis with Panoptic Layout Generation

wb-finalking/PLGAN CVPR 2022

In particular, the stuff layouts can take amorphous shapes and fill up the missing regions left out by the instance layouts.

20
04 Mar 2022

High-Resolution Image Synthesis with Latent Diffusion Models

compvis/stable-diffusion CVPR 2022

By decomposing the image formation process into a sequential application of denoising autoencoders, diffusion models (DMs) achieve state-of-the-art synthesis results on image data and beyond.

65,378
20 Dec 2021

AttrLostGAN: Attribute Controlled Image Synthesis from Reconfigurable Layout and Style

stanifrolov/attrlostgan 25 Mar 2021

In this paper, we propose a method for attribute controlled image synthesis from layout which allows to specify the appearance of individual objects without affecting the rest of the image.

7
25 Mar 2021