no code implementations • 13 Jul 2023 • Moab Arar, Rinon Gal, Yuval Atzmon, Gal Chechik, Daniel Cohen-Or, Ariel Shamir, Amit H. Bermano
Text-to-image (T2I) personalization allows users to guide the creative image generation process by combining their own visual concepts in natural language prompts.
no code implementations • 2 May 2023 • Yoad Tewel, Rinon Gal, Gal Chechik, Yuval Atzmon
The task of T2I personalization poses multiple hard challenges, such as maintaining high visual fidelity while allowing creative control, combining multiple personalized concepts in a single image, and keeping a small model size.
no code implementations • 23 Feb 2023 • Rinon Gal, Moab Arar, Yuval Atzmon, Amit H. Bermano, Gal Chechik, Daniel Cohen-Or
Specifically, we employ two components: First, an encoder that takes as an input a single image of a target concept from a given domain, e. g. a specific face, and learns to map it into a word-embedding representing the concept.
6 code implementations • 2 Aug 2022 • Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H. Bermano, Gal Chechik, Daniel Cohen-Or
Yet, it is unclear how such freedom can be exercised to generate images of specific unique concepts, modify their appearance, or compose them in new roles and novel scenes.
2 code implementations • 4 Apr 2022 • Niv Cohen, Rinon Gal, Eli A. Meirom, Gal Chechik, Yuval Atzmon
We propose an architecture for solving PerVL that operates by extending the input vocabulary of a pretrained model with new word embeddings for the new personalized concepts.
Ranked #7 on
Zero-Shot Composed Image Retrieval (ZS-CIR)
on CIRCO
no code implementations • 2 Feb 2022 • Yuval Atzmon, Eli A. Meirom, Shie Mannor, Gal Chechik
Reasoning and interacting with dynamic environments is a fundamental problem in AI, but it becomes extremely challenging when actions can trigger cascades of cross-dependent events.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Tzuf Paz-Argaman, Yuval Atzmon, Gal Chechik, Reut Tsarfaty
Specifically, given birds' images with free-text descriptions of their species, we learn to classify images of previously-unseen species based on specie descriptions.
1 code implementation • NeurIPS 2020 • Yuval Atzmon, Felix Kreuk, Uri Shalit, Gal Chechik
This leads to consistent misclassification of samples from a new distribution, like new combinations of known components.
1 code implementation • 5 Apr 2020 • Dvir Samuel, Yuval Atzmon, Gal Chechik
Real-world data is predominantly unbalanced and long-tailed, but deep models struggle to recognize rare classes in the presence of frequent classes.
Ranked #1 on
Long-tail learning with class descriptors
on CUB-LT
no code implementations • 26 Jul 2019 • Gilad Vered, Gal Oren, Yuval Atzmon, Gal Chechik
Second, we show that the generated descriptions can be kept close to natural by constraining them to be similar to human descriptions.
no code implementations • 10 Jun 2019 • Roman Visotsky, Yuval Atzmon, Gal Chechik
Here we describe a new approach to learn with fewer samples, by using additional information that is provided per sample.
no code implementations • CVPR 2019 • Yuval Atzmon, Gal Chechik
Specifically, our model consists of three classifiers: A "gating" model that makes soft decisions if a sample is from a "seen" class, and two experts: a ZSL expert, and an expert model for seen classes.
1 code implementation • 7 Jun 2018 • Yuval Atzmon, Gal Chechik
The soft group structure can be learned from data jointly as part of the model, and can also readily incorporate prior knowledge about groups if available.
no code implementations • 27 Aug 2016 • Yuval Atzmon, Jonathan Berant, Vahid Kezami, Amir Globerson, Gal Chechik
Recurrent neural networks have recently been used for learning to describe images using natural language.