text-guided-image-editing
24 papers with code • 0 benchmarks • 0 datasets
Editing images using text prompts.
Benchmarks
These leaderboards are used to track progress in text-guided-image-editing
Most implemented papers
Versatile Diffusion: Text, Images and Variations All in One Diffusion Model
In this work, we expand the existing single-flow diffusion pipeline into a multi-task multimodal network, dubbed Versatile Diffusion (VD), that handles multiple flows of text-to-image, image-to-text, and variations in one unified model.
On Distillation of Guided Diffusion Models
For standard diffusion models trained on the pixel-space, our approach is able to generate images visually comparable to that of the original model using as few as 4 sampling steps on ImageNet 64x64 and CIFAR-10, achieving FID/IS scores comparable to that of the original model while being up to 256 times faster to sample from.
EDICT: Exact Diffusion Inversion via Coupled Transformations
EDICT enables mathematically exact inversion of real and model-generated images by maintaining two coupled noise vectors which are used to invert each other in an alternating fashion.
MDP: A Generalized Framework for Text-Guided Image Editing by Manipulating the Diffusion Path
Image generation using diffusion can be controlled in multiple ways.
ImagenHub: Standardizing the evaluation of conditional image generation models
Recently, a myriad of conditional image generation and editing models have been developed to serve different downstream tasks, including text-to-image generation, text-guided image editing, subject-driven image generation, control-guided image generation, etc.
Blended Diffusion for Text-driven Editing of Natural Images
Natural language offers a highly intuitive interface for image editing.
DE-Net: Dynamic Text-guided Image Editing Adversarial Networks
To solve these limitations, we propose: (i) a Dynamic Editing Block (DEBlock) which composes different editing modules dynamically for various editing requirements.
Blended Latent Diffusion
Our solution leverages a recent text-to-image Latent Diffusion Model (LDM), which speeds up diffusion by operating in a lower-dimensional latent space.
SinDDM: A Single Image Denoising Diffusion Model
Here, we introduce a framework for training a DDM on a single image.
In-Context Learning Unlocked for Diffusion Models
We present Prompt Diffusion, a framework for enabling in-context learning in diffusion-based generative models.