Search Results for author: Daniel Cohen-Or

Found 144 papers, 75 papers with code

Cross-Domain Cascaded Deep Translation

no code implementations ECCV 2020 Oren Katzir, Dani Lischinski, Daniel Cohen-Or

We mitigate this by descending the deep layers of a pre-trained network, where the deep features contain more semantics, and applying the translation between these deep features.

Image-to-Image Translation Translation

Lazy Diffusion Transformer for Interactive Image Editing

no code implementations18 Apr 2024 Yotam Nitzan, Zongze Wu, Richard Zhang, Eli Shechtman, Daniel Cohen-Or, Taesung Park, Michaël Gharbi

We demonstrate that our approach is competitive with state-of-the-art inpainting methods in terms of quality and fidelity while providing a 10x speedup for typical user interactions, where the editing mask represents 10% of the image.

Dynamic Typography: Bringing Text to Life via Video Diffusion Prior

no code implementations17 Apr 2024 Zichen Liu, Yihao Meng, Hao Ouyang, Yue Yu, Bolin Zhao, Daniel Cohen-Or, Huamin Qu

Through quantitative and qualitative evaluations, we demonstrate the effectiveness of our framework in generating coherent text animations that faithfully interpret user prompts while maintaining readability.

Vector Graphics

LCM-Lookahead for Encoder-based Text-to-Image Personalization

no code implementations4 Apr 2024 Rinon Gal, Or Lichter, Elad Richardson, Or Patashnik, Amit H. Bermano, Gal Chechik, Daniel Cohen-Or

In this work, we explore the potential of using such shortcut-mechanisms to guide the personalization of text-to-image models to specific facial identities.

Denoising

Be Yourself: Bounded Attention for Multi-Subject Text-to-Image Generation

no code implementations25 Mar 2024 Omer Dahary, Or Patashnik, Kfir Aberman, Daniel Cohen-Or

Text-to-image diffusion models have an unprecedented ability to generate diverse and high-quality images.

Denoising Text-to-Image Generation

Implicit Style-Content Separation using B-LoRA

no code implementations21 Mar 2024 Yarden Frenkel, Yael Vinker, Ariel Shamir, Daniel Cohen-Or

In this paper, we introduce B-LoRA, a method that leverages LoRA (Low-Rank Adaptation) to implicitly separate the style and content components of a single image, facilitating various image stylization tasks.

Image Stylization

MyVLM: Personalizing VLMs for User-Specific Queries

no code implementations21 Mar 2024 Yuval Alaluf, Elad Richardson, Sergey Tulyakov, Kfir Aberman, Daniel Cohen-Or

To effectively recognize a variety of user-specific concepts, we augment the VLM with external concept heads that function as toggles for the model, enabling the VLM to identify the presence of specific target concepts in a given image.

Image Captioning Language Modelling +2

ReNoise: Real Image Inversion Through Iterative Noising

no code implementations21 Mar 2024 Daniel Garibi, Or Patashnik, Andrey Voynov, Hadar Averbuch-Elor, Daniel Cohen-Or

However, applying these methods to real images necessitates the inversion of the images into the domain of the pretrained diffusion model.

Denoising Image Manipulation

Consolidating Attention Features for Multi-view Image Editing

no code implementations22 Feb 2024 Or Patashnik, Rinon Gal, Daniel Cohen-Or, Jun-Yan Zhu, Fernando de la Torre

In this work, we focus on spatial control-based geometric manipulations and introduce a method to consolidate the editing process across various views.

Generating Non-Stationary Textures using Self-Rectification

1 code implementation5 Jan 2024 Yang Zhou, Rongjun Xiao, Dani Lischinski, Daniel Cohen-Or, Hui Huang

This paper addresses the challenge of example-based non-stationary texture synthesis.

Texture Synthesis

SAI3D: Segment Any Instance in 3D Scenes

no code implementations17 Dec 2023 Yingda Yin, Yuzheng Liu, Yang Xiao, Daniel Cohen-Or, Jingwei Huang, Baoquan Chen

Advancements in 3D instance segmentation have traditionally been tethered to the availability of annotated datasets, limiting their application to a narrow spectrum of object categories.

3D Instance Segmentation Scene Parsing +2

Mismatch Quest: Visual and Textual Feedback for Image-Text Misalignment

no code implementations5 Dec 2023 Brian Gordon, Yonatan Bitton, Yonatan Shafir, Roopal Garg, Xi Chen, Dani Lischinski, Daniel Cohen-Or, Idan Szpektor

While existing image-text alignment models reach high quality binary assessments, they fall short of pinpointing the exact source of misalignment.

Explanation Generation Visual Grounding

Style Aligned Image Generation via Shared Attention

1 code implementation4 Dec 2023 Amir Hertz, Andrey Voynov, Shlomi Fruchter, Daniel Cohen-Or

Large-scale Text-to-Image (T2I) models have rapidly gained prominence across creative fields, generating visually compelling outputs from textual prompts.

Image Generation

AnyLens: A Generative Diffusion Model with Any Rendering Lens

no code implementations29 Nov 2023 Andrey Voynov, Amir Hertz, Moab Arar, Shlomi Fruchter, Daniel Cohen-Or

State-of-the-art diffusion models can generate highly realistic images based on various conditioning like text, segmentation, and depth.

Text Segmentation

CLiC: Concept Learning in Context

no code implementations28 Nov 2023 Mehdi Safaee, Aryan Mikaeili, Or Patashnik, Daniel Cohen-Or, Ali Mahdavi-Amiri

This paper addresses the challenge of learning a local visual pattern of an object from one image, and generating images depicting objects with that pattern.

Object

Breathing Life Into Sketches Using Text-to-Video Priors

no code implementations21 Nov 2023 Rinon Gal, Yael Vinker, Yuval Alaluf, Amit H. Bermano, Daniel Cohen-Or, Ariel Shamir, Gal Chechik

A sketch is one of the most intuitive and versatile tools humans use to convey their ideas visually.

The Chosen One: Consistent Characters in Text-to-Image Diffusion Models

1 code implementation16 Nov 2023 Omri Avrahami, Amir Hertz, Yael Vinker, Moab Arar, Shlomi Fruchter, Ohad Fried, Daniel Cohen-Or, Dani Lischinski

Our quantitative analysis demonstrates that our method strikes a better balance between prompt alignment and identity consistency compared to the baseline methods, and these findings are reinforced by a user study.

Consistent Character Generation Story Visualization

Cross-Image Attention for Zero-Shot Appearance Transfer

no code implementations6 Nov 2023 Yuval Alaluf, Daniel Garibi, Or Patashnik, Hadar Averbuch-Elor, Daniel Cohen-Or

Recent advancements in text-to-image generative models have demonstrated a remarkable ability to capture a deep semantic understanding of images.

Denoising

Noise-Free Score Distillation

no code implementations26 Oct 2023 Oren Katzir, Or Patashnik, Daniel Cohen-Or, Dani Lischinski

Score Distillation Sampling (SDS) has emerged as the de facto approach for text-to-content generation in non-image domains.

MAS: Multi-view Ancestral Sampling for 3D motion generation using 2D diffusion

no code implementations23 Oct 2023 Roy Kapon, Guy Tevet, Daniel Cohen-Or, Amit H. Bermano

We introduce Multi-view Ancestral Sampling (MAS), a method for 3D motion generation, using 2D diffusion models that were trained on motions obtained from in-the-wild videos.

Denoising

Domain-Agnostic Tuning-Encoder for Fast Personalization of Text-To-Image Models

no code implementations13 Jul 2023 Moab Arar, Rinon Gal, Yuval Atzmon, Gal Chechik, Daniel Cohen-Or, Ariel Shamir, Amit H. Bermano

Text-to-image (T2I) personalization allows users to guide the creative image generation process by combining their own visual concepts in natural language prompts.

Image Generation

Facial Reenactment Through a Personalized Generator

no code implementations12 Jul 2023 Ariel Elazary, Yotam Nitzan, Daniel Cohen-Or

In this paper, we propose a novel method for facial reenactment using a personalized generator.

Hallucination

SVNR: Spatially-variant Noise Removal with Denoising Diffusion

no code implementations28 Jun 2023 Naama Pearl, Yaron Brodsky, Dana Berman, Assaf Zomet, Alex Rav Acha, Daniel Cohen-Or, Dani Lischinski

Our formulation also accounts for the correlation that exists between the condition image and the samples along the modified diffusion process.

Image Denoising

SENS: Part-Aware Sketch-based Implicit Neural Shape Modeling

no code implementations9 Jun 2023 Alexandre Binninger, Amir Hertz, Olga Sorkine-Hornung, Daniel Cohen-Or, Raja Giryes

We present SENS, a novel method for generating and editing 3D models from hand-drawn sketches, including those of abstract nature.

Concept Decomposition for Visual Exploration and Inspiration

no code implementations29 May 2023 Yael Vinker, Andrey Voynov, Daniel Cohen-Or, Ariel Shamir

Each node in the tree represents a sub-concept using a learned vector embedding injected into the latent space of a pretrained text-to-image model.

Break-A-Scene: Extracting Multiple Concepts from a Single Image

1 code implementation25 May 2023 Omri Avrahami, Kfir Aberman, Ohad Fried, Daniel Cohen-Or, Dani Lischinski

Text-to-image model personalization aims to introduce a user-provided concept to the model, allowing its synthesis in diverse contexts.

Complex Scene Breaking and Synthesis

A Neural Space-Time Representation for Text-to-Image Personalization

1 code implementation24 May 2023 Yuval Alaluf, Elad Richardson, Gal Metzer, Daniel Cohen-Or

We observe that one can significantly improve the convergence and visual fidelity of the concept by introducing a textual bypass, where our neural mapper additionally outputs a residual that is added to the output of the text encoder.

Denoising

Delta Denoising Score

no code implementations ICCV 2023 Amir Hertz, Kfir Aberman, Daniel Cohen-Or

We introduce Delta Denoising Score (DDS), a novel scoring function for text-based image editing that guides minimal modifications of an input image towards the content described in a target prompt.

Denoising Image-to-Image Translation +2

Set-the-Scene: Global-Local Training for Generating Controllable NeRF Scenes

1 code implementation23 Mar 2023 Dana Cohen-Bar, Elad Richardson, Gal Metzer, Raja Giryes, Daniel Cohen-Or

We show that using proxies allows a wide variety of editing options, such as adjusting the placement of each independent object, removing objects from a scene, or refining an object.

Image Generation Object +1

Localizing Object-level Shape Variations with Text-to-Image Diffusion Models

1 code implementation ICCV 2023 Or Patashnik, Daniel Garibi, Idan Azuri, Hadar Averbuch-Elor, Daniel Cohen-Or

In this paper, we present a technique to generate a collection of images that depicts variations in the shape of a specific object, enabling an object-level shape exploration process.

Denoising Object +1

SKED: Sketch-guided Text-based 3D Editing

no code implementations ICCV 2023 Aryan Mikaeili, Or Perel, Mehdi Safaee, Daniel Cohen-Or, Ali Mahdavi-Amiri

To ensure the generated output adheres to the provided sketches, we propose novel loss functions to generate the desired edits while preserving the density and radiance of the base instance.

Text to 3D

P+: Extended Textual Conditioning in Text-to-Image Generation

no code implementations16 Mar 2023 Andrey Voynov, Qinghao Chu, Daniel Cohen-Or, Kfir Aberman

Furthermore, we utilize the unique properties of this space to achieve previously unattainable results in object-style mixing using text-to-image models.

Denoising Text-to-Image Generation

Word-As-Image for Semantic Typography

no code implementations3 Mar 2023 Shir Iluz, Yael Vinker, Amir Hertz, Daniel Berio, Daniel Cohen-Or, Ariel Shamir

A word-as-image is a semantic typography technique where a word illustration presents a visualization of the meaning of the word, while also preserving its readability.

Encoder-based Domain Tuning for Fast Personalization of Text-to-Image Models

no code implementations23 Feb 2023 Rinon Gal, Moab Arar, Yuval Atzmon, Amit H. Bermano, Gal Chechik, Daniel Cohen-Or

Specifically, we employ two components: First, an encoder that takes as an input a single image of a target concept from a given domain, e. g. a specific face, and learns to map it into a word-embedding representing the concept.

Novel Concepts

Single Motion Diffusion

1 code implementation12 Feb 2023 Sigal Raab, Inbal Leibovitch, Guy Tevet, Moab Arar, Amit H. Bermano, Daniel Cohen-Or

We harness the power of diffusion models and present a denoising network explicitly designed for the task of learning from a single input motion.

Denoising Style Transfer

TEXTure: Text-Guided Texturing of 3D Shapes

1 code implementation3 Feb 2023 Elad Richardson, Gal Metzer, Yuval Alaluf, Raja Giryes, Daniel Cohen-Or

In this paper, we present TEXTure, a novel method for text-guided generation, editing, and transfer of textures for 3D shapes.

Image Generation text-guided-generation

Attend-and-Excite: Attention-Based Semantic Guidance for Text-to-Image Diffusion Models

2 code implementations31 Jan 2023 Hila Chefer, Yuval Alaluf, Yael Vinker, Lior Wolf, Daniel Cohen-Or

Recent text-to-image generative models have demonstrated an unparalleled ability to generate diverse and creative imagery guided by a target text prompt.

Generative Semantic Nursing

CLIPascene: Scene Sketching with Different Types and Levels of Abstraction

no code implementations ICCV 2023 Yael Vinker, Yuval Alaluf, Daniel Cohen-Or, Ariel Shamir

In this paper, we present a method for converting a given scene image into a sketch using different types and multiple levels of abstraction.

Disentanglement

Sketch-Guided Text-to-Image Diffusion Models

no code implementations24 Nov 2022 Andrey Voynov, Kfir Aberman, Daniel Cohen-Or

In this work, we introduce a universal approach to guide a pretrained text-to-image diffusion model, with a spatial map from another domain (e. g., sketch) during inference time.

Denoising Sketch-to-Image Translation

Null-text Inversion for Editing Real Images using Guided Diffusion Models

4 code implementations CVPR 2023 Ron Mokady, Amir Hertz, Kfir Aberman, Yael Pritch, Daniel Cohen-Or

Our Null-text inversion, based on the publicly available Stable Diffusion model, is extensively evaluated on a variety of images and prompt editing, showing high-fidelity editing of real images.

Image Generation Text-based Image Editing

Human Motion Diffusion Model

1 code implementation29 Sep 2022 Guy Tevet, Sigal Raab, Brian Gordon, Yonatan Shafir, Daniel Cohen-Or, Amit H. Bermano

In this paper, we introduce Motion Diffusion Model (MDM), a carefully adapted classifier-free diffusion-based generative model for the human motion domain.

Motion Synthesis

An Image is Worth One Word: Personalizing Text-to-Image Generation using Textual Inversion

7 code implementations2 Aug 2022 Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H. Bermano, Gal Chechik, Daniel Cohen-Or

Yet, it is unclear how such freedom can be exercised to generate images of specific unique concepts, modify their appearance, or compose them in new roles and novel scenes.

Text-to-Image Generation

Prompt-to-Prompt Image Editing with Cross Attention Control

7 code implementations2 Aug 2022 Amir Hertz, Ron Mokady, Jay Tenenbaum, Kfir Aberman, Yael Pritch, Daniel Cohen-Or

Editing is challenging for these generative models, since an innate property of an editing technique is to preserve most of the original image, while in the text-based models, even a small modification of the text prompt often leads to a completely different outcome.

Image Generation Text-based Image Editing

MoDi: Unconditional Motion Synthesis from Diverse Data

1 code implementation CVPR 2023 Sigal Raab, Inbal Leibovitch, Peizhuo Li, Kfir Aberman, Olga Sorkine-Hornung, Daniel Cohen-Or

In this work, we present MoDi -- a generative model trained in an unsupervised setting from an extremely diverse, unstructured and unlabeled dataset.

Motion Interpolation Motion Synthesis

Shape-Pose Disentanglement using SE(3)-equivariant Vector Neurons

no code implementations3 Apr 2022 Oren Katzir, Dani Lischinski, Daniel Cohen-Or

We introduce an unsupervised technique for encoding point clouds into a canonical shape representation, by disentangling shape and pose.

Disentanglement Translation

MyStyle: A Personalized Generative Prior

no code implementations31 Mar 2022 Yotam Nitzan, Kfir Aberman, Qiurui He, Orly Liba, Michal Yarom, Yossi Gandelsman, Inbar Mosseri, Yael Pritch, Daniel Cohen-Or

Given a small reference set of portrait images of a person (~100), we tune the weights of a pretrained StyleGAN face generator to form a local, low-dimensional, personalized manifold in the latent space.

Image Enhancement Super-Resolution

MotionCLIP: Exposing Human Motion Generation to CLIP Space

1 code implementation15 Mar 2022 Guy Tevet, Brian Gordon, Amir Hertz, Amit H. Bermano, Daniel Cohen-Or

MotionCLIP gains its unique power by aligning its latent space with that of the Contrastive Language-Image Pre-training (CLIP) model.

Disentanglement Motion Interpolation

State-of-the-Art in the Architecture, Methods and Applications of StyleGAN

no code implementations28 Feb 2022 Amit H. Bermano, Rinon Gal, Yuval Alaluf, Ron Mokady, Yotam Nitzan, Omer Tov, Or Patashnik, Daniel Cohen-Or

Of these, StyleGAN offers a fascinating case study, owing to its remarkable visual quality and an ability to support a large array of downstream tasks.

Image Generation

Self-Distilled StyleGAN: Towards Generation from Internet Photos

2 code implementations24 Feb 2022 Ron Mokady, Michal Yarom, Omer Tov, Oran Lang, Daniel Cohen-Or, Tali Dekel, Michal Irani, Inbar Mosseri

To meet these challenges, we proposed a StyleGAN-based self-distillation approach, which consists of two main components: (i) A generative-based self-filtering of the dataset to eliminate outlier images, in order to generate an adequate training set, and (ii) Perceptual clustering of the generated images to detect the inherent data modalities, which are then employed to improve StyleGAN's "truncation trick" in the image synthesis process.

Image Generation

Multi-level Latent Space Structuring for Generative Control

no code implementations11 Feb 2022 Oren Katzir, Vicky Perepelook, Dani Lischinski, Daniel Cohen-Or

Truncation is widely used in generative models for improving the quality of the generated samples, at the expense of reducing their diversity.

Self-Conditioned Generative Adversarial Networks for Image Editing

1 code implementation8 Feb 2022 Yunzhe Liu, Rinon Gal, Amit H. Bermano, Baoquan Chen, Daniel Cohen-Or

We compare our models to a wide range of latent editing methods, and show that by alleviating the bias they achieve finer semantic control and better identity preservation through a wider range of transformations.

Fairness

FEAT: Face Editing with Attention

no code implementations6 Feb 2022 Xianxu Hou, Linlin Shen, Or Patashnik, Daniel Cohen-Or, Hui Huang

In this paper, we build on the StyleGAN generator, and present a method that explicitly encourages face manipulation to focus on the intended regions by incorporating learned attention maps.

Disentanglement

Third Time's the Charm? Image and Video Editing with StyleGAN3

1 code implementation31 Jan 2022 Yuval Alaluf, Or Patashnik, Zongze Wu, Asif Zamir, Eli Shechtman, Dani Lischinski, Daniel Cohen-Or

In particular, we demonstrate that while StyleGAN3 can be trained on unaligned data, one can still use aligned data for training, without hindering the ability to generate unaligned imagery.

Disentanglement Image Generation +1

SPAGHETTI: Editing Implicit Shapes Through Part Aware Generation

1 code implementation31 Jan 2022 Amir Hertz, Or Perel, Raja Giryes, Olga Sorkine-Hornung, Daniel Cohen-Or

Neural implicit fields are quickly emerging as an attractive representation for learning based techniques.

3D Shape Modeling

ShapeFormer: Transformer-based Shape Completion via Sparse Representation

1 code implementation CVPR 2022 Xingguang Yan, Liqiang Lin, Niloy J. Mitra, Dani Lischinski, Daniel Cohen-Or, Hui Huang

We present ShapeFormer, a transformer-based network that produces a distribution of object completions, conditioned on incomplete, and possibly noisy, point clouds.

Stitch it in Time: GAN-Based Facial Editing of Real Videos

1 code implementation20 Jan 2022 Rotem Tzaban, Ron Mokady, Rinon Gal, Amit H. Bermano, Daniel Cohen-Or

The ability of Generative Adversarial Networks to encode rich semantics within their latent space has been widely adopted for facial image editing.

Facial Editing

NeuralMLS: Geometry-Aware Control Point Deformation

1 code implementation5 Jan 2022 Meitar Shechter, Rana Hanocka, Gal Metzer, Raja Giryes, Daniel Cohen-Or

In this work, we opt to learn the weighting function, by training a neural network on the control points from a single input shape, and exploit the innate smoothness of neural networks.

Rhythm is a Dancer: Music-Driven Motion Synthesis with Global Structure

no code implementations23 Nov 2021 Andreas Aristidou, Anastasios Yiannakidis, Kfir Aberman, Daniel Cohen-Or, Ariel Shamir, Yiorgos Chrysanthou

In this work, we present a music-driven motion synthesis framework that generates long-term sequences of human motions which are synchronized with the input beats, and jointly form a global structure that respects a specific dance genre.

Motion Synthesis

Mesh Draping: Parametrization-Free Neural Mesh Transfer

no code implementations11 Oct 2021 Amir Hertz, Or Perel, Raja Giryes, Olga Sorkine-Hornung, Daniel Cohen-Or

The method drapes the source mesh over the target geometry and at the same time seeks to preserve the carefully designed characteristics of the source mesh.

StyleGAN-NADA: CLIP-Guided Domain Adaptation of Image Generators

3 code implementations2 Aug 2021 Rinon Gal, Or Patashnik, Haggai Maron, Gal Chechik, Daniel Cohen-Or

Can a generative model be trained to produce images from a specific domain, guided by a text prompt only, without seeing any image?

Domain Adaptation Image Manipulation

LARGE: Latent-Based Regression through GAN Semantics

1 code implementation CVPR 2022 Yotam Nitzan, Rinon Gal, Ofir Brenner, Daniel Cohen-Or

For modern generative frameworks, this semantic encoding manifests as smooth, linear directions which affect image attributes in a disentangled manner.

Attribute regression

StyleFusion: A Generative Model for Disentangling Spatial Segments

1 code implementation15 Jul 2021 Omer Kafri, Or Patashnik, Yuval Alaluf, Daniel Cohen-Or

Inserting the resulting style code into a pre-trained StyleGAN generator results in a single harmonized image in which each semantic region is controlled by one of the input latent codes.

Disentanglement

JOKR: Joint Keypoint Representation for Unsupervised Cross-Domain Motion Retargeting

1 code implementation17 Jun 2021 Ron Mokady, Rotem Tzaban, Sagie Benaim, Amit H. Bermano, Daniel Cohen-Or

To alleviate this problem, we introduce JOKR - a JOint Keypoint Representation that captures the motion common to both the source and target videos, without requiring any object prior or data collection.

Disentanglement motion retargeting

Pivotal Tuning for Latent-based Editing of Real Images

3 code implementations10 Jun 2021 Daniel Roich, Ron Mokady, Amit H. Bermano, Daniel Cohen-Or

The key idea is pivotal tuning - a brief training process that preserves the editing quality of an in-domain latent region, while changing its portrayed identity and appearance.

Facial Editing Image Manipulation

MoCo-Flow: Neural Motion Consensus Flow for Dynamic Humans in Stationary Monocular Cameras

no code implementations8 Jun 2021 Xuelin Chen, Weiyu Li, Daniel Cohen-Or, Niloy J. Mitra, Baoquan Chen

In this paper, we introduce Neural Motion Consensus Flow (MoCo-Flow), a representation that models dynamic humans in stationary monocular cameras using a 4D continuous time-variant function.

Consistent Two-Flow Network for Tele-Registration of Point Clouds

1 code implementation1 Jun 2021 Zihao Yan, Zimu Yi, Ruizhen Hu, Niloy J. Mitra, Daniel Cohen-Or, Hui Huang

In this paper, we present a learning-based technique that alleviates this problem, and allows registration between point clouds, presented in arbitrary poses, and having little or even no overlap, a setting that has been referred to as tele-registration.

Vocal Bursts Valence Prediction

FLEX: Extrinsic Parameters-free Multi-view 3D Human Motion Reconstruction

1 code implementation5 May 2021 Brian Gordon, Sigal Raab, Guy Azov, Raja Giryes, Daniel Cohen-Or

We compare our model to state-of-the-art methods that are not ep-free and show that in the absence of camera parameters, we outperform them by a large margin while obtaining comparable results when camera parameters are available.

3D Human Pose Estimation

Orienting Point Clouds with Dipole Propagation

1 code implementation4 May 2021 Gal Metzer, Rana Hanocka, Denis Zorin, Raja Giryes, Daniele Panozzo, Daniel Cohen-Or

In the global phase, we propagate the orientation across all coherent patches using a dipole propagation.

SAPE: Spatially-Adaptive Progressive Encoding for Neural Optimization

1 code implementation NeurIPS 2021 Amir Hertz, Or Perel, Raja Giryes, Olga Sorkine-Hornung, Daniel Cohen-Or

Multilayer-perceptrons (MLP) are known to struggle with learning functions of high-frequencies, and in particular cases with wide frequency bands.

Representation Learning

ReStyle: A Residual-Based StyleGAN Encoder via Iterative Refinement

2 code implementations ICCV 2021 Yuval Alaluf, Or Patashnik, Daniel Cohen-Or

Instead of directly predicting the latent code of a given real image using a single pass, the encoder is tasked with predicting a residual with respect to the current estimate of the inverted latent code in a self-correcting manner.

Image Generation Real-to-Cartoon translation

StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery

5 code implementations ICCV 2021 Or Patashnik, Zongze Wu, Eli Shechtman, Daniel Cohen-Or, Dani Lischinski

Inspired by the ability of StyleGAN to generate highly realistic images in a variety of domains, much recent work has focused on understanding how to use the latent spaces of StyleGAN to manipulate generated and real images.

Image Manipulation

Clusterplot: High-dimensional Cluster Visualization

no code implementations4 Mar 2021 Or Malkai, Min Lu, Daniel Cohen-Or

We present Clusterplot, a multi-class high-dimensional data visualization tool designed to visualize cluster-level information offering an intuitive understanding of the cluster inter-relations.

Data Visualization Graphics

SWAGAN: A Style-based Wavelet-driven Generative Model

2 code implementations11 Feb 2021 Rinon Gal, Dana Cohen, Amit Bermano, Daniel Cohen-Or

In recent years, considerable progress has been made in the visual quality of Generative Adversarial Networks (GANs).

Image Generation

Only a Matter of Style: Age Transformation Using a Style-Based Regression Model

2 code implementations4 Feb 2021 Yuval Alaluf, Or Patashnik, Daniel Cohen-Or

In this formulation, our method approaches the continuous aging process as a regression task between the input age and desired target age, providing fine-grained control over the generated image.

Face Age Editing Image Manipulation +2

Designing an Encoder for StyleGAN Image Manipulation

8 code implementations4 Feb 2021 Omer Tov, Yuval Alaluf, Yotam Nitzan, Or Patashnik, Daniel Cohen-Or

We then suggest two principles for designing encoders in a manner that allows one to control the proximity of the inversions to regions that StyleGAN was originally trained on.

Image Manipulation

BalaGAN: Image Translation Between Imbalanced Domains via Cross-Modal Transfer

1 code implementation5 Oct 2020 Or Patashnik, Dov Danon, Hao Zhang, Daniel Cohen-Or

State-of-the-art image-to-image translation methods tend to struggle in an imbalanced domain setting, where one image domain lacks richness and diversity.

Image-to-Image Translation Style Transfer +1

Neural Alignment for Face De-pixelization

no code implementations29 Sep 2020 Maayan Shuvi, Noa Fish, Kfir Aberman, Ariel Shamir, Daniel Cohen-Or

Although simple, our framework synthesizes high-quality face reconstructions, demonstrating that given the statistical prior of a human face, multiple aligned pixelated frames contain sufficient information to reconstruct a high-quality approximation of the original signal.

SketchPatch: Sketch Stylization via Seamless Patch-level Synthesis

1 code implementation4 Sep 2020 Noa Fish, Lilach Perry, Amit Bermano, Daniel Cohen-Or

The paradigm of image-to-image translation is leveraged for the benefit of sketch stylization via transfer of geometric textural details.

Image-to-Image Translation Translation

Object Properties Inferring from and Transfer for Human Interaction Motions

no code implementations20 Aug 2020 Qian Zheng, Weikai Wu, Hanting Pan, Niloy Mitra, Daniel Cohen-Or, Hui Huang

In this paper, we present a fine-grained action recognition method that learns to infer such latent object properties from human interaction motion alone.

Fine-grained Action Recognition Object

Self-Sampling for Neural Point Cloud Consolidation

1 code implementation14 Aug 2020 Gal Metzer, Rana Hanocka, Raja Giryes, Daniel Cohen-Or

We introduce a novel technique for neural point cloud consolidation which learns from only the input point cloud.

Inductive Bias

MRGAN: Multi-Rooted 3D Shape Generation with Unsupervised Part Disentanglement

no code implementations25 Jul 2020 Rinon Gal, Amit Bermano, Hao Zhang, Daniel Cohen-Or

Our network encourages disentangled generation of semantic parts via two key ingredients: a root-mixing training strategy which helps decorrelate the different branches to facilitate disentanglement, and a set of loss terms designed with part disentanglement and shape semantics in mind.

3D Shape Generation Disentanglement

Deep Geometric Texture Synthesis

1 code implementation30 Jun 2020 Amir Hertz, Rana Hanocka, Raja Giryes, Daniel Cohen-Or

Learning and synthesizing on local geometric patches enables a genus-oblivious framework, facilitating texture transfer between shapes of different genus.

Image Generation Texture Synthesis

MotioNet: 3D Human Motion Reconstruction from Monocular Video with Skeleton Consistency

no code implementations22 Jun 2020 Mingyi Shi, Kfir Aberman, Andreas Aristidou, Taku Komura, Dani Lischinski, Daniel Cohen-Or, Baoquan Chen

We introduce MotioNet, a deep neural network that directly reconstructs the motion of a 3D human skeleton from monocular video. While previous methods rely on either rigging or inverse kinematics (IK) to associate a consistent skeleton with temporally coherent joint rotations, our method is the first data-driven approach that directly outputs a kinematic skeleton, which is a complete, commonly used, motion representation.

DO-Conv: Depthwise Over-parameterized Convolutional Layer

1 code implementation22 Jun 2020 Jinming Cao, Yangyan Li, Mingchao Sun, Ying Chen, Dani Lischinski, Daniel Cohen-Or, Baoquan Chen, Changhe Tu

Moreover, in the inference phase, the depthwise convolution is folded into the conventional convolution, reducing the computation to be exactly equivalent to that of a convolutional layer without over-parameterization.

Image Classification

Towards a Neural Graphics Pipeline for Controllable Image Generation

no code implementations18 Jun 2020 Xuelin Chen, Daniel Cohen-Or, Baoquan Chen, Niloy J. Mitra

NGP decomposes the image into a set of interpretable appearance feature maps, uncovering direct control handles for controllable image generation.

Image Generation Neural Rendering

Point2Mesh: A Self-Prior for Deformable Meshes

2 code implementations22 May 2020 Rana Hanocka, Gal Metzer, Raja Giryes, Daniel Cohen-Or

We optimize the network weights to deform an initial mesh to shrink-wrap a single input point cloud.

Face Identity Disentanglement via Latent Space Mapping

3 code implementations15 May 2020 Yotam Nitzan, Amit Bermano, Yangyan Li, Daniel Cohen-Or

Learning disentangled representations of data is a fundamental problem in artificial intelligence.

De-identification Disentanglement

Unpaired Motion Style Transfer from Video to Animation

1 code implementation12 May 2020 Kfir Aberman, Yijia Weng, Dani Lischinski, Daniel Cohen-Or, Baoquan Chen

In this paper, we present a novel data-driven framework for motion style transfer, which learns from an unpaired collection of motions with style labels, and enables transferring motion styles not observed during training.

3D Reconstruction Motion Style Transfer +1

Skeleton-Aware Networks for Deep Motion Retargeting

1 code implementation12 May 2020 Kfir Aberman, Peizhuo Li, Dani Lischinski, Olga Sorkine-Hornung, Daniel Cohen-Or, Baoquan Chen

In other words, our operators form the building blocks of a new deep motion processing framework that embeds the motion into a common latent space, shared by a collection of homeomorphic skeletons.

motion retargeting Motion Synthesis

Image Morphing with Perceptual Constraints and STN Alignment

1 code implementation29 Apr 2020 Noa Fish, Richard Zhang, Lilach Perry, Daniel Cohen-Or, Eli Shechtman, Connelly Barnes

In image morphing, a sequence of plausible frames are synthesized and composited together to form a smooth transformation between given instances.

Image Morphing

Single Pair Cross-Modality Super Resolution

no code implementations CVPR 2021 Guy Shacht, Sharon Fogel, Dov Danon, Daniel Cohen-Or, Ilya Leizerson

The network is trained on the two input images only, learns their internal statistics and correlations, and applies them to up-sample the target modality.

Super-Resolution

Structural-analogy from a Single Image Pair

1 code implementation5 Apr 2020 Sagie Benaim, Ron Mokady, Amit Bermano, Daniel Cohen-Or, Lior Wolf

In this paper, we explore the capabilities of neural networks to understand image structure given only a single pair of images, A and B.

Translation Unsupervised Image-To-Image Translation

PointGMM: a Neural GMM Network for Point Clouds

1 code implementation CVPR 2020 Amir Hertz, Rana Hanocka, Raja Giryes, Daniel Cohen-Or

We present PointGMM, a neural network that learns to generate hGMMs which are characteristic of the shape class, and also coincide with the input point cloud.

Unsupervised Multi-Modal Image Registration via Geometry Preserving Image-to-Image Translation

1 code implementation CVPR 2020 Moab Arar, Yiftach Ginger, Dov Danon, Ilya Leizerson, Amit Bermano, Daniel Cohen-Or

In this work, we bypass the difficulties of developing cross-modality similarity measures, by training an image-to-image translation network on the two input modalities.

Autonomous Driving Image Registration +2

A Rotation-Invariant Framework for Deep Point Cloud Analysis

1 code implementation16 Mar 2020 Xianzhi Li, Ruihui Li, Guangyong Chen, Chi-Wing Fu, Daniel Cohen-Or, Pheng-Ann Heng

Recently, many deep neural networks were designed to process 3D point clouds, but a common drawback is that rotation invariance is not ensured, leading to poor generalization to arbitrary orientations.

Point Cloud Generation Retrieval

GANHopper: Multi-Hop GAN for Unsupervised Image-to-Image Translation

1 code implementation ECCV 2020 Wallace Lira, Johannes Merz, Daniel Ritchie, Daniel Cohen-Or, Hao Zhang

Instead of executing translation directly, we steer the translation by requiring the network to produce in-between images that resemble weighted hybrids between images from the input domains.

Translation Unsupervised Image-To-Image Translation

Unsupervised multi-modal Styled Content Generation

no code implementations10 Jan 2020 Omry Sendik, Dani Lischinski, Daniel Cohen-Or

The emergence of deep generative models has recently enabled the automatic generation of massive amounts of graphical content, both in 2D and in 3D.

Cross-Domain Cascaded Deep Feature Translation

no code implementations4 Jun 2019 Oren Katzir, Dani Lischinski, Daniel Cohen-Or

Our translation is performed in a cascaded, deep-to-shallow, fashion, along the deep feature hierarchy: we first translate between the deepest layers that encode the higher-level semantic content of the image, proceeding to translate the shallower layers, conditioned on the deeper ones.

Image-to-Image Translation Translation

Learning Character-Agnostic Motion for Motion Retargeting in 2D

2 code implementations5 May 2019 Kfir Aberman, Rundi Wu, Dani Lischinski, Baoquan Chen, Daniel Cohen-Or

In order to achieve our goal, we learn to extract, directly from a video, a high-level latent motion representation, which is invariant to the skeleton geometry and the camera view.

3D Reconstruction motion retargeting +2

Image Resizing by Reconstruction from Deep Features

no code implementations17 Apr 2019 Moab Arar, Dov Danon, Daniel Cohen-Or, Ariel Shamir

In this paper we perform image resizing in feature space where the deep layers of a neural network contain rich important semantic information.

Implicit Pairs for Boosting Unpaired Image-to-Image Translation

no code implementations15 Apr 2019 Yiftach Ginger, Dov Danon, Hadar Averbuch-Elor, Daniel Cohen-Or

As a result, in recent years more attention has been given to techniques that learn the mapping from unpaired sets.

Image-to-Image Translation Translation

Blind Visual Motif Removal from a Single Image

1 code implementation CVPR 2019 Amir Hertz, Sharon Fogel, Rana Hanocka, Raja Giryes, Daniel Cohen-Or

Many images shared over the web include overlaid objects, or visual motifs, such as text, symbols or drawings, which add a description or decoration to the image.

LOGAN: Unpaired Shape Transform in Latent Overcomplete Space

no code implementations25 Mar 2019 Kangxue Yin, Zhiqin Chen, Hui Huang, Daniel Cohen-Or, Hao Zhang

Our network consists of an autoencoder to encode shapes from the two input domains into a common latent space, where the latent codes concatenate multi-scale shape features, resulting in an overcomplete representation.

Generative Adversarial Network Translation

CrossNet: Latent Cross-Consistency for Unpaired Image Translation

no code implementations14 Jan 2019 Omry Sendik, Dani Lischinski, Daniel Cohen-Or

Recent GAN-based architectures have been able to deliver impressive performance on the general task of image-to-image translation.

Image-to-Image Translation Translation

Deep motifs and motion signatures

1 code implementation ACM Transactions on Graphics 2018 Andreas Aristidou, Daniel Cohen-Or, Jessica K. Hodgins, Yiorgos Chrysanthou, Ariel Shamir

In this paper we introduce motion motifs and motion signatures that are a succinct but descriptive representation of motion sequences.

Descriptive

CompoNet: Learning to Generate the Unseen by Part Synthesis and Composition

1 code implementation ICCV 2019 Nadav Schor, Oren Katzir, Hao Zhang, Daniel Cohen-Or

Data-driven generative modeling has made remarkable progress by leveraging the power of deep neural networks.

Generative Low-Shot Network Expansion

no code implementations ICLR 2018 Adi Hayat, Mark Kliger, Shachar Fleishman, Daniel Cohen-Or

We present a simple yet powerful hard distillation method where the base network is augmented with additional weights to classify the novel classes, while keeping the weights of the base network unchanged.

MeshCNN: A Network with an Edge

1 code implementation16 Sep 2018 Rana Hanocka, Amir Hertz, Noa Fish, Raja Giryes, Shachar Fleishman, Daniel Cohen-Or

In this paper, we utilize the unique properties of the mesh for a direct analysis of 3D shapes using MeshCNN, a convolutional neural network designed specifically for triangular meshes.

3D Part Segmentation Cube Engraving Classification

Multi-Scale Context Intertwining for Semantic Segmentation

no code implementations ECCV 2018 Di Lin, Yuanfeng Ji, Dani Lischinski, Daniel Cohen-Or, Hui Huang

Accurate semantic image segmentation requires the joint consideration of local appearance, semantic information, and global scene context.

Image Segmentation Segmentation +1

Deep Video-Based Performance Cloning

no code implementations21 Aug 2018 Kfir Aberman, Mingyi Shi, Jing Liao, Dani Lischinski, Baoquan Chen, Daniel Cohen-Or

After training a deep generative network using a reference video capturing the appearance and dynamics of a target actor, we are able to generate videos where this actor reenacts other performances.

Structure-aware Generative Network for 3D-Shape Modeling

1 code implementation12 Aug 2018 Zhijie Wu, Xiang Wang, Di Lin, Dani Lischinski, Daniel Cohen-Or, Hui Huang

The key idea is that during the analysis, the two branches exchange information between them, thereby learning the dependencies between structure and geometry and encoding two augmented features, which are then fused into a single latent code.

Graphics

GRAINS: Generative Recursive Autoencoders for INdoor Scenes

no code implementations24 Jul 2018 Manyi Li, Akshay Gadi Patil, Kai Xu, Siddhartha Chaudhuri, Owais Khan, Ariel Shamir, Changhe Tu, Baoquan Chen, Daniel Cohen-Or, Hao Zhang

We present a generative neural network which enables us to generate plausible 3D indoor scenes in large quantities and varieties, easily and highly efficiently.

Graphics

EC-Net: an Edge-aware Point set Consolidation Network

no code implementations ECCV 2018 Lequan Yu, Xianzhi Li, Chi-Wing Fu, Daniel Cohen-Or, Pheng-Ann Heng

In this paper, we present the first deep learning based edge-aware technique to facilitate the consolidation of point clouds.

Surface Reconstruction

Non-Stationary Texture Synthesis by Adversarial Expansion

1 code implementation11 May 2018 Yang Zhou, Zhen Zhu, Xiang Bai, Dani Lischinski, Daniel Cohen-Or, Hui Huang

We demonstrate that this conceptually simple approach is highly effective for capturing large-scale structures, as well as other non-stationary attributes of the input exemplar.

Generative Adversarial Network Texture Synthesis

Neural Best-Buddies: Sparse Cross-Domain Correspondence

2 code implementations10 May 2018 Kfir Aberman, Jing Liao, Mingyi Shi, Dani Lischinski, Baoquan Chen, Daniel Cohen-Or

Correspondence between images is a fundamental problem in computer vision, with a variety of graphics applications.

Image Morphing

P2P-NET: Bidirectional Point Displacement Net for Shape Transform

no code implementations25 Mar 2018 Kangxue Yin, Hui Huang, Daniel Cohen-Or, Hao Zhang

We introduce P2P-NET, a general-purpose deep neural network which learns geometric transformations between point-based shape representations from two domains, e. g., meso-skeletons and surfaces, partial and complete scans, etc.

Clustering-driven Deep Embedding with Pairwise Constraints

1 code implementation22 Mar 2018 Sharon Fogel, Hadar Averbuch-Elor, Jacov Goldberger, Daniel Cohen-Or

In this paper, we depart from centroid-based models and suggest a new framework, called Clustering-driven deep embedding with PAirwise Constraints (CPAC), for non-parametric clustering using a neural network.

Clustering

Outlier Detection for Robust Multi-dimensional Scaling

no code implementations7 Feb 2018 Leonid Blouvshtein, Daniel Cohen-Or

Multi-dimensional scaling (MDS) plays a central role in data-exploration, dimensionality reduction and visualization.

Dimensionality Reduction Outlier Detection

Neuron-level Selective Context Aggregation for Scene Segmentation

no code implementations22 Nov 2017 Zhenhua Wang, Fanglin Gu, Dani Lischinski, Daniel Cohen-Or, Changhe Tu, Baoquan Chen

Contextual information provides important cues for disambiguating visually similar pixels in scene segmentation.

Scene Segmentation Segmentation

Cascaded Feature Network for Semantic Segmentation of RGB-D Images

no code implementations ICCV 2017 Di Lin, Guangyong Chen, Daniel Cohen-Or, Pheng-Ann Heng, Hui Huang

Our approach is to use the available depth to split the image into layers with common visual characteristic of objects/scenes, or common "scene-resolution".

Semantic Segmentation

Bundle Optimization for Multi-aspect Embedding

no code implementations29 Mar 2017 Qiong Zeng, Baoquan Chen, Yanir Kleiman, Daniel Cohen-Or, Yangyan Li

Understanding semantic similarity among images is the core of a wide range of computer vision applications.

Clustering Image Classification +2

Co-segmentation for Space-Time Co-located Collections

no code implementations31 Jan 2017 Hadar Averbuch-Elor, Johannes Kopf, Tamir Hazan, Daniel Cohen-Or

Thus, to disambiguate what the common foreground object is, we introduce a weakly-supervised technique, where we assume only a small seed, given in the form of a single segmented image.

Object Segmentation

Border-Peeling Clustering

1 code implementation14 Dec 2016 Hadar Averbuch-Elor, Nadav Bar, Daniel Cohen-Or

In this paper, we present a novel non-parametric clustering technique.

Clustering

A Holistic Approach for Data-Driven Object Cutout

no code implementations18 Aug 2016 Huayong Xu, Yangyan Li, Wenzheng Chen, Dani Lischinski, Daniel Cohen-Or, Baoquan Chen

We show that the resulting P-maps may be used to evaluate how likely a rectangle proposal is to contain an instance of the class, and further process good proposals to produce an accurate object cutout mask.

Object

Spherical Embedding of Inlier Silhouette Dissimilarities

no code implementations CVPR 2015 Etai Littwin, Hadar Averbuch-Elor, Daniel Cohen-Or

In this paper, we introduce a spherical embedding technique to position a given set of silhouettes of an object as observed from a set of cameras arbitrarily positioned around the object.

Position

Constraints as Features

no code implementations CVPR 2013 Shmuel Asafi, Daniel Cohen-Or

In this paper, we introduce a new approach to constrained clustering which treats the constraints as features.

Constrained Clustering Image Segmentation +1

Cannot find the paper you are looking for? You can Submit a new open access paper.