1 code implementation • 19 Jun 2023 • Ariel Lapid, Idan Achituve, Lior Bracha, Ethan Fetaya
GD-VDM is based on a two-phase generation process involving generating depth videos followed by a novel diffusion Vid2Vid model that generates a coherent real-world video.
1 code implementation • 5 Jun 2023 • Yochai Yemini, Aviv Shamsian, Lior Bracha, Sharon Gannot, Ethan Fetaya
We then condition a diffusion model on the video and use the extracted text through a classifier-guidance mechanism where a pre-trained ASR serves as the classifier.
no code implementations • 30 May 2023 • Lior Bracha, Eitan Shaar, Aviv Shamsian, Ethan Fetaya, Gal Chechik
Our results highlight the potential of using pre-trained visual-semantic models for generating high-quality contextual descriptions.
1 code implementation • CVPR 2019 • Lior Bracha, Gal Chechik
Capturing the interesting components of an image is a key aspect of image understanding.