no code implementations • ECCV 2020 • Oren Katzir, Dani Lischinski, Daniel Cohen-Or
We mitigate this by descending the deep layers of a pre-trained network, where the deep features contain more semantics, and applying the translation between these deep features.
no code implementations • 20 Mar 2025 • Dana Cohen-Bar, Daniel Cohen-Or, Gal Chechik, Yoni Kasten
As 3D content creation continues to grow, transferring semantic textures between 3D meshes remains a significant challenge in computer graphics.
no code implementations • 13 Mar 2025 • Elad Richardson, Kfir Goldberg, Yuval Alaluf, Daniel Cohen-Or
Advanced generative models excel at synthesizing images but often rely on text-based conditioning.
no code implementations • 27 Feb 2025 • Edo Kadosh, Nir Goren, Or Patashnik, Daniel Garibi, Daniel Cohen-Or
This process has an inherent tradeoff between reconstruction and editability, limiting the editing of challenging images such as highly-detailed ones.
no code implementations • 24 Feb 2025 • Inbar Gat, Sigal Raab, Guy Tevet, Yuval Reshef, Amit H. Bermano, Daniel Cohen-Or
Generating motion for arbitrary skeletons is a longstanding challenge in computer graphics, remaining largely unexplored due to the scarcity of diverse datasets and the irregular nature of the data.
no code implementations • 20 Feb 2025 • Rameen Abdal, Or Patashnik, Ivan Skorokhodov, Willi Menapace, Aliaksandr Siarohin, Sergey Tulyakov, Daniel Cohen-Or, Kfir Aberman
In this paper, we introduce Set-and-Sequence, a novel framework for personalizing Diffusion Transformers (DiTs)-based generative video models with dynamic concepts.
no code implementations • 19 Feb 2025 • Sara Dorfman, Dana Cohen-Bar, Rinon Gal, Daniel Cohen-Or
Through comprehensive evaluation, we show that our approach enables more precise control over a larger range of visual concept compositions.
no code implementations • 12 Feb 2025 • Ellie Arar, Yarden Frenkel, Daniel Cohen-Or, Ariel Shamir, Yael Vinker
In this work, we introduce SwiftSketch, a diffusion model for image-conditioned vector sketch generation that can produce high-quality sketches in less than a second.
no code implementations • 7 Jan 2025 • Sagi Polaczek, Yuval Alaluf, Elad Richardson, Yael Vinker, Daniel Cohen-Or
We additionally demonstrate that utilizing a neural representation provides an added benefit of inference-time control, enabling users to dynamically adapt the generated SVG based on user-provided inputs, all with a single learned representation.
no code implementations • 2 Jan 2025 • Or Patashnik, Rinon Gal, Daniil Ostashev, Sergey Tulyakov, Kfir Aberman, Daniel Cohen-Or
In this work, we introduce Nested Attention, a novel mechanism that injects a rich and expressive image representation into the model's existing cross-attention layers.
no code implementations • 2 Jan 2025 • Gaurav Parmar, Or Patashnik, Kuan-Chieh Wang, Daniil Ostashev, Srinivasa Narasimhan, Jun-Yan Zhu, Daniel Cohen-Or, Kfir Aberman
A key challenge in this task is to preserve the identity of the objects depicted in the input visual prompts, while also generating diverse compositions across different images.
no code implementations • 12 Dec 2024 • Guocheng Qian, Kuan-Chieh Wang, Or Patashnik, Negin Heravi, Daniil Ostashev, Sergey Tulyakov, Daniel Cohen-Or, Kfir Aberman
Our approach uses a few-to-many identity reconstruction training paradigm, where a limited set of input images is used to reconstruct multiple target images of the same individual in various poses and expressions.
1 code implementation • 21 Nov 2024 • Omri Avrahami, Or Patashnik, Ohad Fried, Egor Nemchinov, Kfir Aberman, Dani Lischinski, Daniel Cohen-Or
The main challenge is that, unlike the UNet-based models, DiT lacks a coarse-to-fine synthesis structure, making it unclear in which layers to perform the injection.
no code implementations • 2 Oct 2024 • Rinon Gal, Adi Haviv, Yuval Alaluf, Amit H. Bermano, Daniel Cohen-Or, Gal Chechik
Both approaches lead to improved image quality when compared to monolithic models or generic, prompt-independent workflows.
no code implementations • 11 Jun 2024 • Zhengzhe Liu, Qing Liu, Chirui Chang, Jianming Zhang, Daniil Pakhomov, Haitian Zheng, Zhe Lin, Daniel Cohen-Or, Chi-Wing Fu
Deoccluding the hidden portions of objects in a scene is a formidable task, particularly when addressing real-world scenes.
1 code implementation • 10 Jun 2024 • Sigal Raab, Inbar Gat, Nathan Sala, Guy Tevet, Rotem Shalev-Arkushin, Ohad Fried, Amit H. Bermano, Daniel Cohen-Or
Given the remarkable results of motion synthesis with diffusion models, a natural question arises: how can we effectively leverage these models for motion editing?
no code implementations • 8 Jun 2024 • Zhenyu Wang, Jianxi Huang, Zhida Sun, Yuanhao Gong, Daniel Cohen-Or, Min Lu
This work presents a progressive image vectorization technique that reconstructs the raster image as layer-wise vectors from semantic-aligned macro structures to finer details.
1 code implementation • 7 Jun 2024 • Yilin Liu, Jiale Chen, Shanshan Pan, Daniel Cohen-Or, Hao Zhang, Hui Huang
We design a neural network to predict the Voronoi diagram from an input point cloud or distance field via a binary classification.
no code implementations • 3 Jun 2024 • Elad Richardson, Yuval Alaluf, Ali Mahdavi-Amiri, Daniel Cohen-Or
To harness this potential, we introduce pOps, a framework that trains specific semantic operators directly on CLIP image embeddings.
no code implementations • 21 May 2024 • Jingyuan Yang, Jiawei Feng, Weibin Luo, Dani Lischinski, Daniel Cohen-Or, Hui Huang
Affective Image Manipulation (AIM) seeks to modify user-provided images to evoke specific emotional responses.
no code implementations • 18 Apr 2024 • Yotam Nitzan, Zongze Wu, Richard Zhang, Eli Shechtman, Daniel Cohen-Or, Taesung Park, Michaël Gharbi
We demonstrate that our approach is competitive with state-of-the-art inpainting methods in terms of quality and fidelity while providing a 10x speedup for typical user interactions, where the editing mask represents 10% of the image.
no code implementations • 17 Apr 2024 • Zichen Liu, Yihao Meng, Hao Ouyang, Yue Yu, Bolin Zhao, Daniel Cohen-Or, Huamin Qu
Through quantitative and qualitative evaluations, we demonstrate the effectiveness of our framework in generating coherent text animations that faithfully interpret user prompts while maintaining readability.
no code implementations • 4 Apr 2024 • Rinon Gal, Or Lichter, Elad Richardson, Or Patashnik, Amit H. Bermano, Gal Chechik, Daniel Cohen-Or
In this work, we explore the potential of using such shortcut-mechanisms to guide the personalization of text-to-image models to specific facial identities.
no code implementations • 25 Mar 2024 • Omer Dahary, Or Patashnik, Kfir Aberman, Daniel Cohen-Or
Text-to-image diffusion models have an unprecedented ability to generate diverse and high-quality images.
no code implementations • 21 Mar 2024 • Yuval Alaluf, Elad Richardson, Sergey Tulyakov, Kfir Aberman, Daniel Cohen-Or
To effectively recognize a variety of user-specific concepts, we augment the VLM with external concept heads that function as toggles for the model, enabling the VLM to identify the presence of specific target concepts in a given image.
1 code implementation • 21 Mar 2024 • Yarden Frenkel, Yael Vinker, Ariel Shamir, Daniel Cohen-Or
In this paper, we introduce B-LoRA, a method that leverages LoRA (Low-Rank Adaptation) to implicitly separate the style and content components of a single image, facilitating various image stylization tasks.
1 code implementation • 21 Mar 2024 • Daniel Garibi, Or Patashnik, Andrey Voynov, Hadar Averbuch-Elor, Daniel Cohen-Or
However, applying these methods to real images necessitates the inversion of the images into the domain of the pretrained diffusion model.
no code implementations • 22 Feb 2024 • Or Patashnik, Rinon Gal, Daniel Cohen-Or, Jun-Yan Zhu, Fernando de la Torre
In this work, we focus on spatial control-based geometric manipulations and introduce a method to consolidate the editing process across various views.
no code implementations • 11 Jan 2024 • Moab Arar, Andrey Voynov, Amir Hertz, Omri Avrahami, Shlomi Fruchter, Yael Pritch, Daniel Cohen-Or, Ariel Shamir
We term our approach prompt-aligned personalization.
1 code implementation • CVPR 2024 • Yang Zhou, Rongjun Xiao, Dani Lischinski, Daniel Cohen-Or, Hui Huang
This paper addresses the challenge of example-based non-stationary texture synthesis.
no code implementations • CVPR 2024 • Yingda Yin, Yuzheng Liu, Yang Xiao, Daniel Cohen-Or, Jingwei Huang, Baoquan Chen
Advancements in 3D instance segmentation have traditionally been tethered to the availability of annotated datasets, limiting their application to a narrow spectrum of object categories.
1 code implementation • 5 Dec 2023 • Brian Gordon, Yonatan Bitton, Yonatan Shafir, Roopal Garg, Xi Chen, Dani Lischinski, Daniel Cohen-Or, Idan Szpektor
While existing image-text alignment models reach high quality binary assessments, they fall short of pinpointing the exact source of misalignment.
2 code implementations • CVPR 2024 • Amir Hertz, Andrey Voynov, Shlomi Fruchter, Daniel Cohen-Or
Large-scale Text-to-Image (T2I) models have rapidly gained prominence across creative fields, generating visually compelling outputs from textual prompts.
no code implementations • 29 Nov 2023 • Andrey Voynov, Amir Hertz, Moab Arar, Shlomi Fruchter, Daniel Cohen-Or
State-of-the-art diffusion models can generate highly realistic images based on various conditioning like text, segmentation, and depth.
no code implementations • CVPR 2024 • Mehdi Safaee, Aryan Mikaeili, Or Patashnik, Daniel Cohen-Or, Ali Mahdavi-Amiri
This paper addresses the challenge of learning a local visual pattern of an object from one image, and generating images depicting objects with that pattern.
no code implementations • CVPR 2024 • Rinon Gal, Yael Vinker, Yuval Alaluf, Amit H. Bermano, Daniel Cohen-Or, Ariel Shamir, Gal Chechik
A sketch is one of the most intuitive and versatile tools humans use to convey their ideas visually.
1 code implementation • 16 Nov 2023 • Omri Avrahami, Amir Hertz, Yael Vinker, Moab Arar, Shlomi Fruchter, Ohad Fried, Daniel Cohen-Or, Dani Lischinski
Recent advances in text-to-image generation models have unlocked vast potential for visual creativity.
no code implementations • 6 Nov 2023 • Yuval Alaluf, Daniel Garibi, Or Patashnik, Hadar Averbuch-Elor, Daniel Cohen-Or
Recent advancements in text-to-image generative models have demonstrated a remarkable ability to capture a deep semantic understanding of images.
1 code implementation • 3 Nov 2023 • Zhengzhe Liu, Jingyu Hu, Ka-Hei Hui, Xiaojuan Qi, Daniel Cohen-Or, Chi-Wing Fu
This paper presents a new text-guided technique for generating 3D shapes.
no code implementations • 26 Oct 2023 • Oren Katzir, Or Patashnik, Daniel Cohen-Or, Dani Lischinski
Score Distillation Sampling (SDS) has emerged as the de facto approach for text-to-content generation in non-image domains.
no code implementations • CVPR 2024 • Roy Kapon, Guy Tevet, Daniel Cohen-Or, Amit H. Bermano
We introduce Multi-view Ancestral Sampling (MAS), a method for 3D motion generation, using 2D diffusion models that were trained on motions obtained from in-the-wild videos.
1 code implementation • 3 Aug 2023 • Elad Richardson, Kfir Goldberg, Yuval Alaluf, Daniel Cohen-Or
Recent text-to-image generative models have enabled us to transform our words into vibrant, captivating imagery.
no code implementations • ICCV 2023 • Jingyuan Yang, Qirui Huang, Tingting Ding, Dani Lischinski, Daniel Cohen-Or, Hui Huang
Most of the existing work in this area focuses on feature design, while little attention has been paid to dataset construction.
no code implementations • 13 Jul 2023 • Moab Arar, Rinon Gal, Yuval Atzmon, Gal Chechik, Daniel Cohen-Or, Ariel Shamir, Amit H. Bermano
Text-to-image (T2I) personalization allows users to guide the creative image generation process by combining their own visual concepts in natural language prompts.
no code implementations • 12 Jul 2023 • Ariel Elazary, Yotam Nitzan, Daniel Cohen-Or
In this paper, we propose a novel method for facial reenactment using a personalized generator.
no code implementations • 28 Jun 2023 • Naama Pearl, Yaron Brodsky, Dana Berman, Assaf Zomet, Alex Rav Acha, Daniel Cohen-Or, Dani Lischinski
Our formulation also accounts for the correlation that exists between the condition image and the samples along the modified diffusion process.
1 code implementation • 9 Jun 2023 • Alexandre Binninger, Amir Hertz, Olga Sorkine-Hornung, Daniel Cohen-Or, Raja Giryes
We present SENS, a novel method for generating and editing 3D models from hand-drawn sketches, including those of abstract nature.
no code implementations • 29 May 2023 • Yael Vinker, Andrey Voynov, Daniel Cohen-Or, Ariel Shamir
Each node in the tree represents a sub-concept using a learned vector embedding injected into the latent space of a pretrained text-to-image model.
1 code implementation • 25 May 2023 • Omri Avrahami, Kfir Aberman, Ohad Fried, Daniel Cohen-Or, Dani Lischinski
Text-to-image model personalization aims to introduce a user-provided concept to the model, allowing its synthesis in diverse contexts.
1 code implementation • 24 May 2023 • Yuval Alaluf, Elad Richardson, Gal Metzer, Daniel Cohen-Or
We observe that one can significantly improve the convergence and visual fidelity of the concept by introducing a textual bypass, where our neural mapper additionally outputs a residual that is added to the output of the text encoder.
no code implementations • ICCV 2023 • Amir Hertz, Kfir Aberman, Daniel Cohen-Or
We introduce Delta Denoising Score (DDS), a novel scoring function for text-based image editing that guides minimal modifications of an input image towards the content described in a target prompt.
1 code implementation • 23 Mar 2023 • Dana Cohen-Bar, Elad Richardson, Gal Metzer, Raja Giryes, Daniel Cohen-Or
We show that using proxies allows a wide variety of editing options, such as adjusting the placement of each independent object, removing objects from a scene, or refining an object.
1 code implementation • ICCV 2023 • Or Patashnik, Daniel Garibi, Idan Azuri, Hadar Averbuch-Elor, Daniel Cohen-Or
In this paper, we present a technique to generate a collection of images that depicts variations in the shape of a specific object, enabling an object-level shape exploration process.
no code implementations • ICCV 2023 • Aryan Mikaeili, Or Perel, Mehdi Safaee, Daniel Cohen-Or, Ali Mahdavi-Amiri
To ensure the generated output adheres to the provided sketches, we propose novel loss functions to generate the desired edits while preserving the density and radiance of the base instance.
1 code implementation • 16 Mar 2023 • Andrey Voynov, Qinghao Chu, Daniel Cohen-Or, Kfir Aberman
Furthermore, we utilize the unique properties of this space to achieve previously unattainable results in object-style mixing using text-to-image models.
no code implementations • 3 Mar 2023 • Shir Iluz, Yael Vinker, Amir Hertz, Daniel Berio, Daniel Cohen-Or, Ariel Shamir
A word-as-image is a semantic typography technique where a word illustration presents a visualization of the meaning of the word, while also preserving its readability.
no code implementations • 23 Feb 2023 • Rinon Gal, Moab Arar, Yuval Atzmon, Amit H. Bermano, Gal Chechik, Daniel Cohen-Or
Specifically, we employ two components: First, an encoder that takes as an input a single image of a target concept from a given domain, e. g. a specific face, and learns to map it into a word-embedding representing the concept.
1 code implementation • 20 Feb 2023 • Roy Hachnochi, Mingrui Zhao, Nadav Orzech, Rinon Gal, Ali Mahdavi-Amiri, Daniel Cohen-Or, Amit Haim Bermano
Diffusion models have enabled high-quality, conditional image editing capabilities.
1 code implementation • 12 Feb 2023 • Sigal Raab, Inbal Leibovitch, Guy Tevet, Moab Arar, Amit H. Bermano, Daniel Cohen-Or
We harness the power of diffusion models and present a denoising network explicitly designed for the task of learning from a single input motion.
1 code implementation • 3 Feb 2023 • Elad Richardson, Gal Metzer, Yuval Alaluf, Raja Giryes, Daniel Cohen-Or
In this paper, we present TEXTure, a novel method for text-guided generation, editing, and transfer of textures for 3D shapes.
2 code implementations • 31 Jan 2023 • Hila Chefer, Yuval Alaluf, Yael Vinker, Lior Wolf, Daniel Cohen-Or
Recent text-to-image generative models have demonstrated an unparalleled ability to generate diverse and creative imagery guided by a target text prompt.
2 code implementations • CVPR 2023 • Yotam Nitzan, Michaël Gharbi, Richard Zhang, Taesung Park, Jun-Yan Zhu, Daniel Cohen-Or, Eli Shechtman
First, we note the generator contains a meaningful, pretrained latent space.
no code implementations • ICCV 2023 • Yael Vinker, Yuval Alaluf, Daniel Cohen-Or, Ariel Shamir
In this paper, we present a method for converting a given scene image into a sketch using different types and multiple levels of abstraction.
no code implementations • 24 Nov 2022 • Andrey Voynov, Kfir Aberman, Daniel Cohen-Or
In this work, we introduce a universal approach to guide a pretrained text-to-image diffusion model, with a spatial map from another domain (e. g., sketch) during inference time.
4 code implementations • CVPR 2023 • Ron Mokady, Amir Hertz, Kfir Aberman, Yael Pritch, Daniel Cohen-Or
Our Null-text inversion, based on the publicly available Stable Diffusion model, is extensively evaluated on a variety of images and prompt editing, showing high-fidelity editing of real images.
Ranked #6 on
Text-based Image Editing
on PIE-Bench
2 code implementations • CVPR 2023 • Gal Metzer, Elad Richardson, Or Patashnik, Raja Giryes, Daniel Cohen-Or
This unique combination of text and shape guidance allows for increased control over the generation process.
Ranked #3 on
Text to 3D
on T$^3$Bench
1 code implementation • 29 Sep 2022 • Guy Tevet, Sigal Raab, Brian Gordon, Yonatan Shafir, Daniel Cohen-Or, Amit H. Bermano
In this paper, we introduce Motion Diffusion Model (MDM), a carefully adapted classifier-free diffusion-based generative model for the human motion domain.
Ranked #1 on
Motion Synthesis
on HumanAct12
8 code implementations • 2 Aug 2022 • Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H. Bermano, Gal Chechik, Daniel Cohen-Or
Yet, it is unclear how such freedom can be exercised to generate images of specific unique concepts, modify their appearance, or compose them in new roles and novel scenes.
Ranked #7 on
Personalized Image Generation
on DreamBooth
7 code implementations • 2 Aug 2022 • Amir Hertz, Ron Mokady, Jay Tenenbaum, Kfir Aberman, Yael Pritch, Daniel Cohen-Or
Editing is challenging for these generative models, since an innate property of an editing technique is to preserve most of the original image, while in the text-based models, even a small modification of the text prompt often leads to a completely different outcome.
Ranked #17 on
Text-based Image Editing
on PIE-Bench
1 code implementation • CVPR 2023 • Sigal Raab, Inbal Leibovitch, Peizhuo Li, Kfir Aberman, Olga Sorkine-Hornung, Daniel Cohen-Or
In this work, we present MoDi -- a generative model trained in an unsupervised setting from an extremely diverse, unstructured and unlabeled dataset.
1 code implementation • 3 Apr 2022 • Oren Katzir, Dani Lischinski, Daniel Cohen-Or
We introduce an unsupervised technique for encoding point clouds into a canonical shape representation, by disentangling shape and pose.
no code implementations • 31 Mar 2022 • Yotam Nitzan, Kfir Aberman, Qiurui He, Orly Liba, Michal Yarom, Yossi Gandelsman, Inbar Mosseri, Yael Pritch, Daniel Cohen-Or
Given a small reference set of portrait images of a person (~100), we tune the weights of a pretrained StyleGAN face generator to form a local, low-dimensional, personalized manifold in the latent space.
1 code implementation • 15 Mar 2022 • Guy Tevet, Brian Gordon, Amir Hertz, Amit H. Bermano, Daniel Cohen-Or
MotionCLIP gains its unique power by aligning its latent space with that of the Contrastive Language-Image Pre-training (CLIP) model.
no code implementations • 28 Feb 2022 • Amit H. Bermano, Rinon Gal, Yuval Alaluf, Ron Mokady, Yotam Nitzan, Omer Tov, Or Patashnik, Daniel Cohen-Or
Of these, StyleGAN offers a fascinating case study, owing to its remarkable visual quality and an ability to support a large array of downstream tasks.
2 code implementations • 24 Feb 2022 • Ron Mokady, Michal Yarom, Omer Tov, Oran Lang, Daniel Cohen-Or, Tali Dekel, Michal Irani, Inbar Mosseri
To meet these challenges, we proposed a StyleGAN-based self-distillation approach, which consists of two main components: (i) A generative-based self-filtering of the dataset to eliminate outlier images, in order to generate an adequate training set, and (ii) Perceptual clustering of the generated images to detect the inherent data modalities, which are then employed to improve StyleGAN's "truncation trick" in the image synthesis process.
no code implementations • 11 Feb 2022 • Oren Katzir, Vicky Perepelook, Dani Lischinski, Daniel Cohen-Or
Truncation is widely used in generative models for improving the quality of the generated samples, at the expense of reducing their diversity.
1 code implementation • 11 Feb 2022 • Yael Vinker, Ehsan Pajouheshgar, Jessica Y. Bo, Roman Christian Bachmann, Amit Haim Bermano, Daniel Cohen-Or, Amir Zamir, Ariel Shamir
Abstraction is at the heart of sketching due to the simple and minimal nature of line drawings.
1 code implementation • 8 Feb 2022 • Yunzhe Liu, Rinon Gal, Amit H. Bermano, Baoquan Chen, Daniel Cohen-Or
We compare our models to a wide range of latent editing methods, and show that by alleviating the bias they achieve finer semantic control and better identity preservation through a wider range of transformations.
no code implementations • 6 Feb 2022 • Xianxu Hou, Linlin Shen, Or Patashnik, Daniel Cohen-Or, Hui Huang
In this paper, we build on the StyleGAN generator, and present a method that explicitly encourages face manipulation to focus on the intended regions by incorporating learned attention maps.
1 code implementation • 31 Jan 2022 • Amir Hertz, Or Perel, Raja Giryes, Olga Sorkine-Hornung, Daniel Cohen-Or
Neural implicit fields are quickly emerging as an attractive representation for learning based techniques.
1 code implementation • 31 Jan 2022 • Yuval Alaluf, Or Patashnik, Zongze Wu, Asif Zamir, Eli Shechtman, Dani Lischinski, Daniel Cohen-Or
In particular, we demonstrate that while StyleGAN3 can be trained on unaligned data, one can still use aligned data for training, without hindering the ability to generate unaligned imagery.
1 code implementation • CVPR 2022 • Xingguang Yan, Liqiang Lin, Niloy J. Mitra, Dani Lischinski, Daniel Cohen-Or, Hui Huang
We present ShapeFormer, a transformer-based network that produces a distribution of object completions, conditioned on incomplete, and possibly noisy, point clouds.
1 code implementation • 20 Jan 2022 • Rotem Tzaban, Ron Mokady, Rinon Gal, Amit H. Bermano, Daniel Cohen-Or
The ability of Generative Adversarial Networks to encode rich semantics within their latent space has been widely adopted for facial image editing.
1 code implementation • 5 Jan 2022 • Meitar Shechter, Rana Hanocka, Gal Metzer, Raja Giryes, Daniel Cohen-Or
In this work, we opt to learn the weighting function, by training a neural network on the control points from a single input shape, and exploit the innate smoothness of neural networks.
no code implementations • 23 Nov 2021 • Andreas Aristidou, Anastasios Yiannakidis, Kfir Aberman, Daniel Cohen-Or, Ariel Shamir, Yiorgos Chrysanthou
In this work, we present a music-driven motion synthesis framework that generates long-term sequences of human motions which are synchronized with the input beats, and jointly form a global structure that respects a specific dance genre.
no code implementations • 11 Oct 2021 • Amir Hertz, Or Perel, Raja Giryes, Olga Sorkine-Hornung, Daniel Cohen-Or
The method drapes the source mesh over the target geometry and at the same time seeks to preserve the carefully designed characteristics of the source mesh.
3 code implementations • 2 Aug 2021 • Rinon Gal, Or Patashnik, Haggai Maron, Gal Chechik, Daniel Cohen-Or
Can a generative model be trained to produce images from a specific domain, guided by a text prompt only, without seeing any image?
1 code implementation • CVPR 2022 • Yotam Nitzan, Rinon Gal, Ofir Brenner, Daniel Cohen-Or
For modern generative frameworks, this semantic encoding manifests as smooth, linear directions which affect image attributes in a disentangled manner.
1 code implementation • 15 Jul 2021 • Omer Kafri, Or Patashnik, Yuval Alaluf, Daniel Cohen-Or
Inserting the resulting style code into a pre-trained StyleGAN generator results in a single harmonized image in which each semantic region is controlled by one of the input latent codes.
1 code implementation • 17 Jun 2021 • Ron Mokady, Rotem Tzaban, Sagie Benaim, Amit H. Bermano, Daniel Cohen-Or
To alleviate this problem, we introduce JOKR - a JOint Keypoint Representation that captures the motion common to both the source and target videos, without requiring any object prior or data collection.
3 code implementations • 10 Jun 2021 • Daniel Roich, Ron Mokady, Amit H. Bermano, Daniel Cohen-Or
The key idea is pivotal tuning - a brief training process that preserves the editing quality of an in-domain latent region, while changing its portrayed identity and appearance.
no code implementations • 8 Jun 2021 • Xuelin Chen, Weiyu Li, Daniel Cohen-Or, Niloy J. Mitra, Baoquan Chen
In this paper, we introduce Neural Motion Consensus Flow (MoCo-Flow), a representation that models dynamic humans in stationary monocular cameras using a 4D continuous time-variant function.
1 code implementation • 1 Jun 2021 • Zihao Yan, Zimu Yi, Ruizhen Hu, Niloy J. Mitra, Daniel Cohen-Or, Hui Huang
In this paper, we present a learning-based technique that alleviates this problem, and allows registration between point clouds, presented in arbitrary poses, and having little or even no overlap, a setting that has been referred to as tele-registration.
1 code implementation • 30 May 2021 • Gal Metzer, Rana Hanocka, Raja Giryes, Niloy J. Mitra, Daniel Cohen-Or
We present a technique for visualizing point clouds using a neural network.
1 code implementation • 5 May 2021 • Brian Gordon, Sigal Raab, Guy Azov, Raja Giryes, Daniel Cohen-Or
We compare our model to state-of-the-art methods that are not ep-free and show that in the absence of camera parameters, we outperform them by a large margin while obtaining comparable results when camera parameters are available.
Ranked #23 on
3D Human Pose Estimation
on Human3.6M
1 code implementation • 4 May 2021 • Gal Metzer, Rana Hanocka, Denis Zorin, Raja Giryes, Daniele Panozzo, Daniel Cohen-Or
In the global phase, we propagate the orientation across all coherent patches using a dipole propagation.
1 code implementation • NeurIPS 2021 • Amir Hertz, Or Perel, Raja Giryes, Olga Sorkine-Hornung, Daniel Cohen-Or
Multilayer-perceptrons (MLP) are known to struggle with learning functions of high-frequencies, and in particular cases with wide frequency bands.
2 code implementations • ICCV 2021 • Yuval Alaluf, Or Patashnik, Daniel Cohen-Or
Instead of directly predicting the latent code of a given real image using a single pass, the encoder is tasked with predicting a residual with respect to the current estimate of the inverted latent code in a self-correcting manner.
5 code implementations • ICCV 2021 • Or Patashnik, Zongze Wu, Eli Shechtman, Daniel Cohen-Or, Dani Lischinski
Inspired by the ability of StyleGAN to generate highly realistic images in a variety of domains, much recent work has focused on understanding how to use the latent spaces of StyleGAN to manipulate generated and real images.
no code implementations • 4 Mar 2021 • Or Malkai, Min Lu, Daniel Cohen-Or
We present Clusterplot, a multi-class high-dimensional data visualization tool designed to visualize cluster-level information offering an intuitive understanding of the cluster inter-relations.
Data Visualization
Graphics
2 code implementations • 11 Feb 2021 • Rinon Gal, Dana Cohen, Amit Bermano, Daniel Cohen-Or
In recent years, considerable progress has been made in the visual quality of Generative Adversarial Networks (GANs).
Ranked #11 on
Image Generation
on FFHQ 1024 x 1024
8 code implementations • 4 Feb 2021 • Omer Tov, Yuval Alaluf, Yotam Nitzan, Or Patashnik, Daniel Cohen-Or
We then suggest two principles for designing encoders in a manner that allows one to control the proximity of the inversions to regions that StyleGAN was originally trained on.
2 code implementations • 4 Feb 2021 • Yuval Alaluf, Or Patashnik, Daniel Cohen-Or
In this formulation, our method approaches the continuous aging process as a regression task between the input age and desired target age, providing fine-grained control over the generated image.
1 code implementation • 5 Oct 2020 • Or Patashnik, Dov Danon, Hao Zhang, Daniel Cohen-Or
State-of-the-art image-to-image translation methods tend to struggle in an imbalanced domain setting, where one image domain lacks richness and diversity.
no code implementations • 29 Sep 2020 • Maayan Shuvi, Noa Fish, Kfir Aberman, Ariel Shamir, Daniel Cohen-Or
Although simple, our framework synthesizes high-quality face reconstructions, demonstrating that given the statistical prior of a human face, multiple aligned pixelated frames contain sufficient information to reconstruct a high-quality approximation of the original signal.
1 code implementation • 4 Sep 2020 • Noa Fish, Lilach Perry, Amit Bermano, Daniel Cohen-Or
The paradigm of image-to-image translation is leveraged for the benefit of sketch stylization via transfer of geometric textural details.
no code implementations • 20 Aug 2020 • Qian Zheng, Weikai Wu, Hanting Pan, Niloy Mitra, Daniel Cohen-Or, Hui Huang
In this paper, we present a fine-grained action recognition method that learns to infer such latent object properties from human interaction motion alone.
1 code implementation • 14 Aug 2020 • Gal Metzer, Rana Hanocka, Raja Giryes, Daniel Cohen-Or
We introduce a novel technique for neural point cloud consolidation which learns from only the input point cloud.
10 code implementations • CVPR 2021 • Elad Richardson, Yuval Alaluf, Or Patashnik, Yotam Nitzan, Yaniv Azar, Stav Shapiro, Daniel Cohen-Or
We present a generic image-to-image translation framework, pixel2style2pixel (pSp).
no code implementations • 25 Jul 2020 • Rinon Gal, Amit Bermano, Hao Zhang, Daniel Cohen-Or
Our network encourages disentangled generation of semantic parts via two key ingredients: a root-mixing training strategy which helps decorrelate the different branches to facilitate disentanglement, and a set of loss terms designed with part disentanglement and shape semantics in mind.
1 code implementation • 30 Jun 2020 • Amir Hertz, Rana Hanocka, Raja Giryes, Daniel Cohen-Or
Learning and synthesizing on local geometric patches enables a genus-oblivious framework, facilitating texture transfer between shapes of different genus.
1 code implementation • 22 Jun 2020 • Jinming Cao, Yangyan Li, Mingchao Sun, Ying Chen, Dani Lischinski, Daniel Cohen-Or, Baoquan Chen, Changhe Tu
Moreover, in the inference phase, the depthwise convolution is folded into the conventional convolution, reducing the computation to be exactly equivalent to that of a convolutional layer without over-parameterization.
no code implementations • 22 Jun 2020 • Mingyi Shi, Kfir Aberman, Andreas Aristidou, Taku Komura, Dani Lischinski, Daniel Cohen-Or, Baoquan Chen
We introduce MotioNet, a deep neural network that directly reconstructs the motion of a 3D human skeleton from monocular video. While previous methods rely on either rigging or inverse kinematics (IK) to associate a consistent skeleton with temporally coherent joint rotations, our method is the first data-driven approach that directly outputs a kinematic skeleton, which is a complete, commonly used, motion representation.
no code implementations • 18 Jun 2020 • Xuelin Chen, Daniel Cohen-Or, Baoquan Chen, Niloy J. Mitra
NGP decomposes the image into a set of interpretable appearance feature maps, uncovering direct control handles for controllable image generation.
2 code implementations • 22 May 2020 • Rana Hanocka, Gal Metzer, Raja Giryes, Daniel Cohen-Or
We optimize the network weights to deform an initial mesh to shrink-wrap a single input point cloud.
3 code implementations • 15 May 2020 • Yotam Nitzan, Amit Bermano, Yangyan Li, Daniel Cohen-Or
Learning disentangled representations of data is a fundamental problem in artificial intelligence.
1 code implementation • 12 May 2020 • Kfir Aberman, Yijia Weng, Dani Lischinski, Daniel Cohen-Or, Baoquan Chen
In this paper, we present a novel data-driven framework for motion style transfer, which learns from an unpaired collection of motions with style labels, and enables transferring motion styles not observed during training.
1 code implementation • 12 May 2020 • Kfir Aberman, Peizhuo Li, Dani Lischinski, Olga Sorkine-Hornung, Daniel Cohen-Or, Baoquan Chen
In other words, our operators form the building blocks of a new deep motion processing framework that embeds the motion into a common latent space, shared by a collection of homeomorphic skeletons.
1 code implementation • 29 Apr 2020 • Noa Fish, Richard Zhang, Lilach Perry, Daniel Cohen-Or, Eli Shechtman, Connelly Barnes
In image morphing, a sequence of plausible frames are synthesized and composited together to form a smooth transformation between given instances.
no code implementations • CVPR 2021 • Guy Shacht, Sharon Fogel, Dov Danon, Daniel Cohen-Or, Ilya Leizerson
The network is trained on the two input images only, learns their internal statistics and correlations, and applies them to up-sample the target modality.
1 code implementation • 5 Apr 2020 • Sagie Benaim, Ron Mokady, Amit Bermano, Daniel Cohen-Or, Lior Wolf
In this paper, we explore the capabilities of neural networks to understand image structure given only a single pair of images, A and B.
1 code implementation • CVPR 2020 • Amir Hertz, Rana Hanocka, Raja Giryes, Daniel Cohen-Or
We present PointGMM, a neural network that learns to generate hGMMs which are characteristic of the shape class, and also coincide with the input point cloud.
1 code implementation • CVPR 2020 • Moab Arar, Yiftach Ginger, Dov Danon, Ilya Leizerson, Amit Bermano, Daniel Cohen-Or
In this work, we bypass the difficulties of developing cross-modality similarity measures, by training an image-to-image translation network on the two input modalities.
1 code implementation • 16 Mar 2020 • Xianzhi Li, Ruihui Li, Guangyong Chen, Chi-Wing Fu, Daniel Cohen-Or, Pheng-Ann Heng
Recently, many deep neural networks were designed to process 3D point clouds, but a common drawback is that rotation invariance is not ensured, leading to poor generalization to arbitrary orientations.
1 code implementation • ECCV 2020 • Wallace Lira, Johannes Merz, Daniel Ritchie, Daniel Cohen-Or, Hao Zhang
Instead of executing translation directly, we steer the translation by requiring the network to produce in-between images that resemble weighted hybrids between images from the input domains.
no code implementations • 10 Jan 2020 • Omry Sendik, Dani Lischinski, Daniel Cohen-Or
The emergence of deep generative models has recently enabled the automatic generation of massive amounts of graphical content, both in 2D and in 3D.
3 code implementations • ICCV 2019 • Ruihui Li, Xianzhi Li, Chi-Wing Fu, Daniel Cohen-Or, Pheng-Ann Heng
Point clouds acquired from range scans are often sparse, noisy, and non-uniform.
no code implementations • 4 Jun 2019 • Oren Katzir, Dani Lischinski, Daniel Cohen-Or
Our translation is performed in a cascaded, deep-to-shallow, fashion, along the deep feature hierarchy: we first translate between the deepest layers that encode the higher-level semantic content of the image, proceeding to translate the shallower layers, conditioned on the deeper ones.
2 code implementations • 5 May 2019 • Kfir Aberman, Rundi Wu, Dani Lischinski, Baoquan Chen, Daniel Cohen-Or
In order to achieve our goal, we learn to extract, directly from a video, a high-level latent motion representation, which is invariant to the skeleton geometry and the camera view.
no code implementations • 5 May 2019 • Xianzhi Li, Lequan Yu, Chi-Wing Fu, Daniel Cohen-Or, Pheng-Ann Heng
This paper presents a novel approach to learn and detect distinctive regions on 3D shapes.
no code implementations • 17 Apr 2019 • Moab Arar, Dov Danon, Daniel Cohen-Or, Ariel Shamir
In this paper we perform image resizing in feature space where the deep layers of a neural network contain rich important semantic information.
no code implementations • 15 Apr 2019 • Yiftach Ginger, Dov Danon, Hadar Averbuch-Elor, Daniel Cohen-Or
As a result, in recent years more attention has been given to techniques that learn the mapping from unpaired sets.
1 code implementation • CVPR 2019 • Amir Hertz, Sharon Fogel, Rana Hanocka, Raja Giryes, Daniel Cohen-Or
Many images shared over the web include overlaid objects, or visual motifs, such as text, symbols or drawings, which add a description or decoration to the image.
no code implementations • 26 Mar 2019 • Felix Petersen, Amit H. Bermano, Oliver Deussen, Daniel Cohen-Or
The long-coveted task of reconstructing 3D geometry from images is still a standing problem.
no code implementations • 25 Mar 2019 • Kangxue Yin, Zhiqin Chen, Hui Huang, Daniel Cohen-Or, Hao Zhang
Our network consists of an autoencoder to encode shapes from the two input domains into a common latent space, where the latent codes concatenate multi-scale shape features, resulting in an overcomplete representation.
no code implementations • 14 Jan 2019 • Omry Sendik, Dani Lischinski, Daniel Cohen-Or
Recent GAN-based architectures have been able to deliver impressive performance on the general task of image-to-image translation.
1 code implementation • ACM Transactions on Graphics (Proc. SIGGRAPH ASIA) 2019 • Hao Wang, Nadav Schor, Ruizhen Hu, Haibin Huang, Daniel Cohen-Or, Hui Huang
We also introduce new means to measure and evaluate the performance of an adversarial generative model.
1 code implementation • ACM Transactions on Graphics 2018 • Andreas Aristidou, Daniel Cohen-Or, Jessica K. Hodgins, Yiorgos Chrysanthou, Ariel Shamir
In this paper we introduce motion motifs and motion signatures that are a succinct but descriptive representation of motion sequences.
3 code implementations • CVPR 2019 • Wang Yifan, Shihao Wu, Hui Huang, Daniel Cohen-Or, Olga Sorkine-Hornung
We present a detail-driven deep neural network for point set upsampling.
1 code implementation • ICCV 2019 • Nadav Schor, Oren Katzir, Hao Zhang, Daniel Cohen-Or
Data-driven generative modeling has made remarkable progress by leveraging the power of deep neural networks.
no code implementations • ICLR 2018 • Adi Hayat, Mark Kliger, Shachar Fleishman, Daniel Cohen-Or
We present a simple yet powerful hard distillation method where the base network is augmented with additional weights to classify the novel classes, while keeping the weights of the base network unchanged.
1 code implementation • 16 Sep 2018 • Rana Hanocka, Amir Hertz, Noa Fish, Raja Giryes, Shachar Fleishman, Daniel Cohen-Or
In this paper, we utilize the unique properties of the mesh for a direct analysis of 3D shapes using MeshCNN, a convolutional neural network designed specifically for triangular meshes.
no code implementations • ECCV 2018 • Di Lin, Yuanfeng Ji, Dani Lischinski, Daniel Cohen-Or, Hui Huang
Accurate semantic image segmentation requires the joint consideration of local appearance, semantic information, and global scene context.
no code implementations • 21 Aug 2018 • Kfir Aberman, Mingyi Shi, Jing Liao, Dani Lischinski, Baoquan Chen, Daniel Cohen-Or
After training a deep generative network using a reference video capturing the appearance and dynamics of a target actor, we are able to generate videos where this actor reenacts other performances.
1 code implementation • 12 Aug 2018 • Zhijie Wu, Xiang Wang, Di Lin, Dani Lischinski, Daniel Cohen-Or, Hui Huang
The key idea is that during the analysis, the two branches exchange information between them, thereby learning the dependencies between structure and geometry and encoding two augmented features, which are then fused into a single latent code.
Graphics
no code implementations • 24 Jul 2018 • Manyi Li, Akshay Gadi Patil, Kai Xu, Siddhartha Chaudhuri, Owais Khan, Ariel Shamir, Changhe Tu, Baoquan Chen, Daniel Cohen-Or, Hao Zhang
We present a generative neural network which enables us to generate plausible 3D indoor scenes in large quantities and varieties, easily and highly efficiently.
Graphics
no code implementations • ECCV 2018 • Lequan Yu, Xianzhi Li, Chi-Wing Fu, Daniel Cohen-Or, Pheng-Ann Heng
In this paper, we present the first deep learning based edge-aware technique to facilitate the consolidation of point clouds.
1 code implementation • 11 May 2018 • Yang Zhou, Zhen Zhu, Xiang Bai, Dani Lischinski, Daniel Cohen-Or, Hui Huang
We demonstrate that this conceptually simple approach is highly effective for capturing large-scale structures, as well as other non-stationary attributes of the input exemplar.
2 code implementations • 10 May 2018 • Kfir Aberman, Jing Liao, Mingyi Shi, Dani Lischinski, Baoquan Chen, Daniel Cohen-Or
Correspondence between images is a fundamental problem in computer vision, with a variety of graphics applications.
1 code implementation • 23 Apr 2018 • Rana Hanocka, Noa Fish, Zhenhua Wang, Raja Giryes, Shachar Fleishman, Daniel Cohen-Or
The process of aligning a pair of shapes is a fundamental operation in computer graphics.
no code implementations • 25 Mar 2018 • Kangxue Yin, Hui Huang, Daniel Cohen-Or, Hao Zhang
We introduce P2P-NET, a general-purpose deep neural network which learns geometric transformations between point-based shape representations from two domains, e. g., meso-skeletons and surfaces, partial and complete scans, etc.
1 code implementation • 22 Mar 2018 • Sharon Fogel, Hadar Averbuch-Elor, Jacov Goldberger, Daniel Cohen-Or
In this paper, we depart from centroid-based models and suggest a new framework, called Clustering-driven deep embedding with PAirwise Constraints (CPAC), for non-parametric clustering using a neural network.
no code implementations • 7 Feb 2018 • Leonid Blouvshtein, Daniel Cohen-Or
Multi-dimensional scaling (MDS) plays a central role in data-exploration, dimensionality reduction and visualization.
3 code implementations • CVPR 2018 • Lequan Yu, Xianzhi Li, Chi-Wing Fu, Daniel Cohen-Or, Pheng-Ann Heng
Learning and analyzing 3D point clouds with deep networks is challenging due to the sparseness and irregularity of the data.
Ranked #3 on
Point Cloud Super Resolution
on SHREC15
no code implementations • 22 Nov 2017 • Zhenhua Wang, Fanglin Gu, Dani Lischinski, Daniel Cohen-Or, Changhe Tu, Baoquan Chen
Contextual information provides important cues for disambiguating visually similar pixels in scene segmentation.
no code implementations • ICCV 2017 • Di Lin, Guangyong Chen, Daniel Cohen-Or, Pheng-Ann Heng, Hui Huang
Our approach is to use the available depth to split the image into layers with common visual characteristic of objects/scenes, or common "scene-resolution".
Ranked #79 on
Semantic Segmentation
on NYU Depth v2
no code implementations • 29 Mar 2017 • Qiong Zeng, Baoquan Chen, Yanir Kleiman, Daniel Cohen-Or, Yangyan Li
Understanding semantic similarity among images is the core of a wide range of computer vision applications.
no code implementations • 31 Jan 2017 • Hadar Averbuch-Elor, Johannes Kopf, Tamir Hazan, Daniel Cohen-Or
Thus, to disambiguate what the common foreground object is, we introduce a weakly-supervised technique, where we assume only a small seed, given in the form of a single segmented image.
1 code implementation • 14 Dec 2016 • Hadar Averbuch-Elor, Nadav Bar, Daniel Cohen-Or
In this paper, we present a novel non-parametric clustering technique.
no code implementations • 18 Aug 2016 • Huayong Xu, Yangyan Li, Wenzheng Chen, Dani Lischinski, Daniel Cohen-Or, Baoquan Chen
We show that the resulting P-maps may be used to evaluate how likely a rectangle proposal is to contain an instance of the class, and further process good proposals to produce an accurate object cutout mask.
no code implementations • 10 Apr 2016 • Wenzheng Chen, Huan Wang, Yangyan Li, Hao Su, Zhenhua Wang, Changhe Tu, Dani Lischinski, Daniel Cohen-Or, Baoquan Chen
Human 3D pose estimation from a single image is a challenging task with numerous applications.
no code implementations • CVPR 2015 • Etai Littwin, Hadar Averbuch-Elor, Daniel Cohen-Or
In this paper, we introduce a spherical embedding technique to position a given set of silhouettes of an object as observed from a set of cameras arbitrarily positioned around the object.
no code implementations • CVPR 2013 • Shmuel Asafi, Daniel Cohen-Or
In this paper, we introduce a new approach to constrained clustering which treats the constraints as features.