no code implementations • 13 Jun 2023 • Allan Jabri, Sjoerd van Steenkiste, Emiel Hoogeboom, Mehdi S. M. Sajjadi, Thomas Kipf
In this paper, we leverage recent progress in diffusion models to equip 3D scene representation learning models with the ability to render high-fidelity novel views, while retaining benefits such as object-level scene editing to a large degree.
1 code implementation • NeurIPS 2023 • Dave Epstein, Allan Jabri, Ben Poole, Alexei A. Efros, Aleksander Holynski
However, many aspects of an image are difficult or impossible to convey through text.
2 code implementations • 22 Dec 2022 • Allan Jabri, David Fleet, Ting Chen
We show how to leverage recurrence by conditioning the latent tokens at each forward pass of the reverse diffusion process with those from prior computation, i. e. latent self-conditioning.
Ranked #1 on
Image Generation
on ImageNet 64x64
1 code implementation • 4 Apr 2022 • Pavel Tokmakov, Allan Jabri, Jie Li, Adrien Gaidon
This paper proposes a self-supervised objective for learning representations that localize objects under occlusion - a property known as object permanence.
1 code implementation • CVPR 2022 • Zhipeng Bao, Pavel Tokmakov, Allan Jabri, Yu-Xiong Wang, Adrien Gaidon, Martial Hebert
Our experiments demonstrate that, despite only capturing a small subset of the objects that move, this signal is enough to generalize to segment both moving and static instances of dynamic objects.
no code implementations • CVPR 2022 • Zhangxing Bian, Allan Jabri, Alexei A. Efros, Andrew Owens
A range of video modeling tasks, from optical flow to multiple object tracking, share the same fundamental challenge: establishing space-time correspondence.
1 code implementation • NeurIPS 2020 • Allan Jabri, Andrew Owens, Alexei A. Efros
We cast correspondence as prediction of links in a space-time graph constructed from video.
1 code implementation • 23 Dec 2019 • Richard Li, Allan Jabri, Trevor Darrell, Pulkit Agrawal
Learning robotic manipulation tasks using reinforcement learning with sparse rewards is currently impractical due to the outrageous data requirements.
no code implementations • NeurIPS 2019 • Allan Jabri, Kyle Hsu, Ben Eysenbach, Abhishek Gupta, Sergey Levine, Chelsea Finn
In experiments on vision-based navigation and manipulation domains, we show that the algorithm allows for unsupervised meta-learning that transfers to downstream tasks specified by hand-crafted reward functions and serves as pre-training for more efficient supervised meta-learning of test task distributions.
1 code implementation • CVPR 2019 • Xiaolong Wang, Allan Jabri, Alexei A. Efros
We introduce a self-supervised method for learning visual correspondence from unlabeled video.
1 code implementation • ICML 2018 • Aravind Srinivas, Allan Jabri, Pieter Abbeel, Sergey Levine, Chelsea Finn
A key challenge in complex visuomotor control is learning abstract representations that are effective for specifying goals, planning, and generalization.
1 code implementation • 2 Apr 2018 • Aravind Srinivas, Allan Jabri, Pieter Abbeel, Sergey Levine, Chelsea Finn
We find that the representations learned are not only effective for goal-directed visual imitation via gradient-based trajectory optimization, but can also provide a metric for specifying goals using images.
no code implementations • NAACL 2018 • Douwe Kiela, Alexis Conneau, Allan Jabri, Maximilian Nickel
We introduce a variety of models, trained on a supervised image captioning corpus to predict the image features for a given caption, to perform sentence representation grounding.
no code implementations • 31 Jan 2017 • Marco Baroni, Armand Joulin, Allan Jabri, Germàn Kruszewski, Angeliki Lazaridou, Klemen Simonic, Tomas Mikolov
With machine learning successfully applied to new daunting problems almost every day, general AI starts looking like an attainable goal.
no code implementations • ICCV 2017 • Ang Li, Allan Jabri, Armand Joulin, Laurens van der Maaten
Real-world image recognition systems need to recognize tens of thousands of classes that constitute a plethora of visual concepts.
Ranked #2 on
Zero-Shot Transfer Image Classification
on SUN
3 code implementations • 27 Jun 2016 • Allan Jabri, Armand Joulin, Laurens van der Maaten
Visual question answering (VQA) is an interesting learning setting for evaluating the abilities and shortcomings of current systems for image understanding.
no code implementations • 6 Nov 2015 • Armand Joulin, Laurens van der Maaten, Allan Jabri, Nicolas Vasilache
We train convolutional networks on a dataset of 100 million Flickr photos and captions, and show that these networks produce features that perform well in a range of vision problems.