Search Results for author: Tali Dekel

Found 36 papers, 16 papers with code

MultiDiffusion: Fusing Diffusion Paths for Controlled Image Generation

2 code implementations16 Feb 2023 Omer Bar-Tal, Lior Yariv, Yaron Lipman, Tali Dekel

In this work, we present MultiDiffusion, a unified framework that enables versatile and controllable image generation, using a pre-trained text-to-image diffusion model, without any further training or finetuning.

Text-to-Image Generation

TokenFlow: Consistent Diffusion Features for Consistent Video Editing

1 code implementation19 Jul 2023 Michal Geyer, Omer Bar-Tal, Shai Bagon, Tali Dekel

In this work, we present a framework that harnesses the power of a text-to-image diffusion model for the task of text-driven video editing.

Video Editing

Text2LIVE: Text-Driven Layered Image and Video Editing

1 code implementation5 Apr 2022 Omer Bar-Tal, Dolev Ofri-Amar, Rafail Fridman, Yoni Kasten, Tali Dekel

Given an input image or video and a target text prompt, our goal is to edit the appearance of existing objects (e. g., object's texture) or augment the scene with visual effects (e. g., smoke, fire) in a semantically meaningful manner.

Video Editing

Plug-and-Play Diffusion Features for Text-Driven Image-to-Image Translation

3 code implementations CVPR 2023 Narek Tumanyan, Michal Geyer, Shai Bagon, Tali Dekel

Large-scale text-to-image generative models have been a revolutionary breakthrough in the evolution of generative AI, allowing us to synthesize diverse images that convey highly complex visual concepts.

Image-to-Image Translation Text-based Image Editing +1

Layered Neural Atlases for Consistent Video Editing

2 code implementations23 Sep 2021 Yoni Kasten, Dolev Ofri, Oliver Wang, Tali Dekel

We present a method that decomposes, or "unwraps", an input video into a set of layered 2D atlases, each providing a unified representation of the appearance of an object (or background) over the video.

Style Transfer Video Editing +2

Splicing ViT Features for Semantic Appearance Transfer

1 code implementation CVPR 2022 Narek Tumanyan, Omer Bar-Tal, Shai Bagon, Tali Dekel

Specifically, our goal is to generate an image in which objects in a source structure image are "painted" with the visual appearance of their semantically related objects in a target appearance image.

Image Generation Style Transfer

Deep ViT Features as Dense Visual Descriptors

1 code implementation10 Dec 2021 Shir Amir, Yossi Gandelsman, Shai Bagon, Tali Dekel

To distill the power of ViT features from convoluted design choices, we restrict ourselves to lightweight zero-shot methodologies (e. g., binning and clustering) applied directly to the features.

Feature Upsampling

Self-Distilled StyleGAN: Towards Generation from Internet Photos

2 code implementations24 Feb 2022 Ron Mokady, Michal Yarom, Omer Tov, Oran Lang, Daniel Cohen-Or, Tali Dekel, Michal Irani, Inbar Mosseri

To meet these challenges, we proposed a StyleGAN-based self-distillation approach, which consists of two main components: (i) A generative-based self-filtering of the dataset to eliminate outlier images, in order to generate an adequate training set, and (ii) Perceptual clustering of the generated images to detect the inherent data modalities, which are then employed to improve StyleGAN's "truncation trick" in the image synthesis process.

Image Generation

Looking to Listen at the Cocktail Party: A Speaker-Independent Audio-Visual Model for Speech Separation

5 code implementations10 Apr 2018 Ariel Ephrat, Inbar Mosseri, Oran Lang, Tali Dekel, Kevin Wilson, Avinatan Hassidim, William T. Freeman, Michael Rubinstein

Solving this task using only audio as input is extremely challenging and does not provide an association of the separated speech signals with speakers in the video.

Speech Separation

Semantic Pyramid for Image Generation

2 code implementations CVPR 2020 Assaf Shocher, Yossi Gandelsman, Inbar Mosseri, Michal Yarom, Michal Irani, William T. Freeman, Tali Dekel

We demonstrate that our model results in a versatile and flexible framework that can be used in various classic and novel image generation tasks.

General Classification Image Generation +2

SpeedNet: Learning the Speediness in Videos

1 code implementation CVPR 2020 Sagie Benaim, Ariel Ephrat, Oran Lang, Inbar Mosseri, William T. Freeman, Michael Rubinstein, Michal Irani, Tali Dekel

We demonstrate how those learned features can boost the performance of self-supervised action recognition, and can be used for video retrieval.

Binary Classification Retrieval +2

Teaching CLIP to Count to Ten

1 code implementation ICCV 2023 Roni Paiss, Ariel Ephrat, Omer Tov, Shiran Zada, Inbar Mosseri, Michal Irani, Tali Dekel

Our counting loss is deployed over automatically-created counterfactual examples, each consisting of an image and a caption containing an incorrect object count.

counterfactual Image Retrieval +4

Layered Neural Rendering for Retiming People in Video

1 code implementation16 Sep 2020 Erika Lu, Forrester Cole, Tali Dekel, Weidi Xie, Andrew Zisserman, David Salesin, William T. Freeman, Michael Rubinstein

We present a method for retiming people in an ordinary, natural video -- manipulating and editing the time in which different motions of individuals in the video occur.

Neural Rendering

MoSculp: Interactive Visualization of Shape and Time

no code implementations14 Sep 2018 Xiuming Zhang, Tali Dekel, Tianfan Xue, Andrew Owens, Qiurui He, Jiajun Wu, Stefanie Mueller, William T. Freeman

We present a system that allows users to visualize complex human motion via 3D motion sculptures---a representation that conveys the 3D structure swept by a human body as it moves through space.

Modifying Non-Local Variations Across Multiple Views

no code implementations CVPR 2018 Tal Tlusty, Tomer Michaeli, Tali Dekel, Lihi Zelnik-Manor

We present an algorithm for modifying small non-local variations between repeating structures and patterns in multiple images of the same scene.

On the Effectiveness of Visible Watermarks

no code implementations CVPR 2017 Tali Dekel, Michael Rubinstein, Ce Liu, William T. Freeman

Since such an attack relies on the consistency of watermarks across image collection, we explore and evaluate how it is affected by various types of inconsistencies in the watermark embedding that could potentially be used to make watermarking more secured.

Image Matting

Omnimatte: Associating Objects and Their Effects in Video

no code implementations CVPR 2021 Erika Lu, Forrester Cole, Tali Dekel, Andrew Zisserman, William T. Freeman, Michael Rubinstein

We show results on real-world videos containing interactions between different types of subjects (cars, animals, people) and complex effects, ranging from semi-transparent elements such as smoke and reflections, to fully opaque effects such as objects attached to the subject.

Consistent Depth of Moving Objects in Video

no code implementations2 Aug 2021 Zhoutong Zhang, Forrester Cole, Richard Tucker, William T. Freeman, Tali Dekel

We present a method to estimate depth of a dynamic scene, containing arbitrary moving objects, from an ordinary video captured with a moving camera.

Depth Estimation Depth Prediction +2

Diverse Video Generation from a Single Video

no code implementations11 May 2022 Niv Haim, Ben Feinstein, Niv Granot, Assaf Shocher, Shai Bagon, Tali Dekel, Michal Irani

GANs are able to perform generation and manipulation tasks, trained on a single video.

Video Generation

Imagic: Text-Based Real Image Editing with Diffusion Models

no code implementations CVPR 2023 Bahjat Kawar, Shiran Zada, Oran Lang, Omer Tov, Huiwen Chang, Tali Dekel, Inbar Mosseri, Michal Irani

In this paper we demonstrate, for the very first time, the ability to apply complex (e. g., non-rigid) text-guided semantic edits to a single real image.

Style Transfer

Neural Congealing: Aligning Images to a Joint Semantic Atlas

no code implementations CVPR 2023 Dolev Ofri-Amar, Michal Geyer, Yoni Kasten, Tali Dekel

We present Neural Congealing -- a zero-shot self-supervised framework for detecting and jointly aligning semantically-common content across a given set of images.

State of the Art on Diffusion Models for Visual Computing

no code implementations11 Oct 2023 Ryan Po, Wang Yifan, Vladislav Golyanik, Kfir Aberman, Jonathan T. Barron, Amit H. Bermano, Eric Ryan Chan, Tali Dekel, Aleksander Holynski, Angjoo Kanazawa, C. Karen Liu, Lingjie Liu, Ben Mildenhall, Matthias Nießner, Björn Ommer, Christian Theobalt, Peter Wonka, Gordon Wetzstein

The field of visual computing is rapidly advancing due to the emergence of generative artificial intelligence (AI), which unlocks unprecedented capabilities for the generation, editing, and reconstruction of images, videos, and 3D scenes.

Disentangling Structure and Appearance in ViT Feature Space

no code implementations20 Nov 2023 Narek Tumanyan, Omer Bar-Tal, Shir Amir, Shai Bagon, Tali Dekel

Specifically, our goal is to generate an image in which objects in a source structure image are "painted" with the visual appearance of their semantically related objects in a target appearance image.

Semantic Segmentation

Space-Time Diffusion Features for Zero-Shot Text-Driven Motion Transfer

no code implementations28 Nov 2023 Danah Yatim, Rafail Fridman, Omer Bar-Tal, Yoni Kasten, Tali Dekel

This loss guides the generation process to preserve the overall motion of the input video while complying with the target object in terms of shape and fine-grained motion traits.

SceneScape: Text-Driven Consistent Scene Generation

no code implementations NeurIPS 2023 Rafail Fridman, Amit Abecasis, Yoni Kasten, Tali Dekel

We present a method for text-driven perpetual view generation -- synthesizing long-term videos of various scenes solely, given an input text prompt describing the scene and camera poses.

Depth Estimation Depth Prediction +3

DINO-Tracker: Taming DINO for Self-Supervised Point Tracking in a Single Video

no code implementations21 Mar 2024 Narek Tumanyan, Assaf Singer, Shai Bagon, Tali Dekel

Specifically, our framework simultaneously adopts DINO's features to fit to the motion observations of the test video, while training a tracker that directly leverages the refined features.

Point Tracking

Cannot find the paper you are looking for? You can Submit a new open access paper.