Search Results for author: Ohad Fried

Found 33 papers, 20 papers with code

REED-VAE: RE-Encode Decode Training for Iterative Image Editing with Diffusion Models

1 code implementation26 Apr 2025 Gal Almog, Ariel Shamir, Ohad Fried

While latent diffusion models achieve impressive image editing results, their application to iterative editing of the same image is severely restricted.

ImageRAG: Dynamic Image Retrieval for Reference-Guided Image Generation

no code implementations13 Feb 2025 Rotem Shalev-Arkushin, Rinon Gal, Amit H. Bermano, Ohad Fried

To address this challenge, we explore the usage of Retrieval-Augmented Generation (RAG) with image generation models.

Image Generation Image Retrieval +3

Tiled Diffusion

no code implementations CVPR 2025 Or Madar, Ohad Fried

This paper presents Tiled Diffusion, a novel approach that extends the capabilities of diffusion models to accommodate the generation of cohesive tiling patterns across various domains of image synthesis that require tiling.

Image Generation single-image-generation

Memories of Forgotten Concepts

no code implementations CVPR 2025 Matan Rusanovsky, Shimon Malnick, Amir Jevnisek, Ohad Fried, Shai Avidan

Diffusion models dominate the space of text-to-image generation, yet they may produce undesirable outputs, including explicit content or private data.

Text to Image Generation Text-to-Image Generation

Stable Flow: Vital Layers for Training-Free Image Editing

1 code implementation CVPR 2025 Omri Avrahami, Or Patashnik, Ohad Fried, Egor Nemchinov, Kfir Aberman, Dani Lischinski, Daniel Cohen-Or

The main challenge is that, unlike the UNet-based models, DiT lacks a coarse-to-fine synthesis structure, making it unclear in which layers to perform the injection.

Text-based Image Editing

Advancing Fine-Grained Classification by Structure and Subject Preserving Augmentation

1 code implementation20 Jun 2024 Eyal Michaeli, Ohad Fried

While these models have been used to generate training data for classification tasks, their effectiveness in full-dataset training of FGVC models remains under-explored.

Data Augmentation Diversity +3

V-LASIK: Consistent Glasses-Removal from Videos Using Synthetic Data

no code implementations20 Jun 2024 Rotem Shalev-Arkushin, Aharon Azulay, Tavi Halperin, Eitan Richardson, Amit H. Bermano, Ohad Fried

We show that despite data imperfection, by learning from our generated data and leveraging the prior of pretrained diffusion models, our model is able to perform the desired edit consistently while preserving the original video content.

Attribute Video Editing

Monkey See, Monkey Do: Harnessing Self-attention in Motion Diffusion for Zero-shot Motion Transfer

1 code implementation10 Jun 2024 Sigal Raab, Inbar Gat, Nathan Sala, Guy Tevet, Rotem Shalev-Arkushin, Ohad Fried, Amit H. Bermano, Daniel Cohen-Or

Given the remarkable results of motion synthesis with diffusion models, a natural question arises: how can we effectively leverage these models for motion editing?

Motion Synthesis Style Transfer

DiffUHaul: A Training-Free Method for Object Dragging in Images

no code implementations3 Jun 2024 Omri Avrahami, Rinon Gal, Gal Chechik, Ohad Fried, Dani Lischinski, Arash Vahdat, Weili Nie

In this work, we propose a training-free method, dubbed DiffUHaul, that harnesses the spatial understanding of a localized text-to-image model, for the object dragging task.

spatial-aware image editing

Diffusing Colors: Image Colorization with Text Guided Diffusion

no code implementations7 Dec 2023 Nir Zabari, Aharon Azulay, Alexey Gorkor, Tavi Halperin, Ohad Fried

To tackle these issues, we present a novel image colorization framework that utilizes image diffusion techniques with granular text prompts.

Colorization Image Colorization

Differential Diffusion: Giving Each Pixel Its Strength

4 code implementations1 Jun 2023 Eran Levin, Ohad Fried

While current techniques enable user control over the degree of change in an image edit, the controllability is limited to global changes over an entire edited region.

Image Generation Text-based Image Editing

Break-A-Scene: Extracting Multiple Concepts from a Single Image

1 code implementation25 May 2023 Omri Avrahami, Kfir Aberman, Ohad Fried, Daniel Cohen-Or, Dani Lischinski

Text-to-image model personalization aims to introduce a user-provided concept to the model, allowing its synthesis in diverse contexts.

Complex Scene Breaking and Synthesis

Deep Image Fingerprint: Towards Low Budget Synthetic Image Detection and Model Lineage Analysis

1 code implementation19 Mar 2023 Sergey Sinitsa, Ohad Fried

Our method can detect images from a known generative model and enable us to establish relationships between fine-tuned generative models.

Inductive Bias Synthetic Image Detection

Prediction of Scene Plausibility

no code implementations2 Dec 2022 Or Nachmias, Ohad Fried, Ariel Shamir

Understanding the 3D world from 2D images involves more than detection and segmentation of the objects within the scene.

Prediction Scene Understanding

FakeOut: Leveraging Out-of-domain Self-supervision for Multi-modal Video Deepfake Detection

1 code implementation1 Dec 2022 Gil Knafo, Ohad Fried

This poses a problem, especially in the era of social media, as synthetic videos of speaking humans can be used to spread misinformation in a convincing manner.

DeepFake Detection Face Swapping +1

Taming Normalizing Flows

2 code implementations29 Nov 2022 Shimon Malnick, Shai Avidan, Ohad Fried

We propose an algorithm for taming Normalizing Flow models - changing the probability that the model will produce a specific image or image category.

Neural Font Rendering

1 code implementation27 Nov 2022 Daniel Anderson, Ariel Shamir, Ohad Fried

Recent advances in deep learning techniques and applications have revolutionized artistic creation and manipulation in many domains (text, images, music); however, fonts have not yet been integrated with deep learning architectures in a manner that supports their multi-scale nature.

Deep Learning

SpaText: Spatio-Textual Representation for Controllable Image Generation

no code implementations CVPR 2023 Omri Avrahami, Thomas Hayes, Oran Gafni, Sonal Gupta, Yaniv Taigman, Devi Parikh, Dani Lischinski, Ohad Fried, Xi Yin

Due to lack of large-scale datasets that have a detailed textual description for each region in the image, we choose to leverage the current large-scale text-to-image datasets and base our approach on a novel CLIP-based spatio-textual representation, and show its effectiveness on two state-of-the-art diffusion models: pixel-based and latent-based.

Text to Image Generation Text-to-Image Generation

GEFF: Improving Any Clothes-Changing Person ReID Model using Gallery Enrichment with Face Features

1 code implementation24 Nov 2022 Daniel Arkushin, Bar Cohen, Shmuel Peleg, Ohad Fried

Combined with the latest ReID models, our method achieves new SOTA results on the PRCC, LTCC, CCVID, LaST and VC-Clothes benchmarks and the proposed 42Street dataset.

Person Re-Identification

Ham2Pose: Animating Sign Language Notation into Pose Sequences

1 code implementation CVPR 2023 Rotem Shalev-Arkushin, Amit Moryossef, Ohad Fried

Additionally, we offer a new distance measurement that considers missing keypoints, to measure the distance between pose sequences using DTW-MJE.

Dynamic Time Warping

Blended Latent Diffusion

1 code implementation6 Jun 2022 Omri Avrahami, Ohad Fried, Dani Lischinski

Our solution leverages a recent text-to-image Latent Diffusion Model (LDM), which speeds up diffusion by operating in a lower-dimensional latent space.

Image Inpainting spatial-aware image editing +2

DDNeRF: Depth Distribution Neural Radiance Fields

1 code implementation30 Mar 2022 David Dadon, Ohad Fried, Yacov Hel-Or

We present depth distribution neural radiance field (DDNeRF), a new method that significantly increases sampling efficiency along rays during training while achieving superior results for a given sampling budget.

NeRF Novel View Synthesis

Disentangled3D: Learning a 3D Generative Model with Disentangled Geometry and Appearance from Monocular Images

no code implementations CVPR 2022 Ayush Tewari, Mallikarjun B R, Xingang Pan, Ohad Fried, Maneesh Agrawala, Christian Theobalt

Our model can disentangle the geometry and appearance variations in the scene, i. e., we can independently sample from the geometry and appearance spaces of the generative model.

Disentanglement

DeepShadow: Neural Shape from Shadow

1 code implementation28 Mar 2022 Asaf Karnieli, Ohad Fried, Yacov Hel-Or

We show that the self and cast shadows not only do not disturb 3D reconstruction, but can be used alone, as a strong learning signal, to recover the depth map and surface normals.

3D Reconstruction

GAN Cocktail: mixing GANs without dataset access

1 code implementation7 Jun 2021 Omri Avrahami, Dani Lischinski, Ohad Fried

In the second stage, we merge the rooted models by averaging their weights and fine-tuning them for each specific domain, using only data generated by the original trained models.

Transfer Learning

Iterative Text-based Editing of Talking-heads Using Neural Retargeting

no code implementations21 Nov 2020 Xinwei Yao, Ohad Fried, Kayvon Fatahalian, Maneesh Agrawala

We present a text-based tool for editing talking-head video that enables an iterative editing workflow.

State of the Art on Neural Rendering

no code implementations8 Apr 2020 Ayush Tewari, Ohad Fried, Justus Thies, Vincent Sitzmann, Stephen Lombardi, Kalyan Sunkavalli, Ricardo Martin-Brualla, Tomas Simon, Jason Saragih, Matthias Nießner, Rohit Pandey, Sean Fanello, Gordon Wetzstein, Jun-Yan Zhu, Christian Theobalt, Maneesh Agrawala, Eli Shechtman, Dan B. Goldman, Michael Zollhöfer

Neural rendering is a new and rapidly emerging field that combines generative machine learning techniques with physical knowledge from computer graphics, e. g., by the integration of differentiable rendering into network training.

BIG-bench Machine Learning Image Generation +2

Lifespan Age Transformation Synthesis

2 code implementations ECCV 2020 Roy Or-El, Soumyadip Sengupta, Ohad Fried, Eli Shechtman, Ira Kemelmacher-Shlizerman

Most existing aging methods are limited to changing the texture, overlooking transformations in head shape that occur during the human aging and growth process.

Face Age Editing Generative Adversarial Network +5

Text-based Editing of Talking-head Video

1 code implementation4 Jun 2019 Ohad Fried, Ayush Tewari, Michael Zollhöfer, Adam Finkelstein, Eli Shechtman, Dan B. Goldman, Kyle Genova, Zeyu Jin, Christian Theobalt, Maneesh Agrawala

To edit a video, the user has to only edit the transcript, and an optimization strategy then chooses segments of the input corpus as base material.

Face Model Sentence +3

Cannot find the paper you are looking for? You can Submit a new open access paper.