no code implementations • 18 Nov 2024 • Apurva Kalia, Dilip Krishnan, Soha Hassoun
Results: We introduce in this paper a novel paradigm (JESTR) for annotation.
1 code implementation • 5 Jan 2024 • Jiawei Yang, Katie Z Luo, Jiefeng Li, Congyue Deng, Leonidas Guibas, Dilip Krishnan, Kilian Q Weinberger, Yonglong Tian, Yue Wang
In the second stage, we train a lightweight transformer block to predict clean features from raw ViT outputs, leveraging the derived estimates of the clean features as supervision.
2 code implementations • CVPR 2024 • Yonglong Tian, Lijie Fan, KaiFeng Chen, Dina Katabi, Dilip Krishnan, Phillip Isola
We introduce SynCLR, a novel approach for learning visual representations exclusively from synthetic images and synthetic captions, without any real data.
1 code implementation • CVPR 2024 • Lijie Fan, KaiFeng Chen, Dilip Krishnan, Dina Katabi, Phillip Isola, Yonglong Tian
Our findings also suggest that scaling synthetic data can be particularly effective in scenarios such as: (1) when there is a limited supply of real images for a supervised problem (e. g., fewer than 0. 5 million images in ImageNet), (2) when the evaluation dataset diverges significantly from the training data, indicating the out-of-distribution scenario, or (3) when synthetic data is used in conjunction with real images, as demonstrated in the training of CLIP models.
no code implementations • 1 Dec 2023 • KaiFeng Chen, Daniel Salz, Huiwen Chang, Kihyuk Sohn, Dilip Krishnan, Mojtaba Seyedhosseini
On K-Nearest-Neighbor image retrieval evaluation with ImageNet-1k, the same model outperforms the baseline by 1. 32%.
no code implementations • 5 Oct 2023 • Tianhong Li, Sangnie Bhardwaj, Yonglong Tian, Han Zhang, Jarred Barber, Dina Katabi, Guillaume Lajoie, Huiwen Chang, Dilip Krishnan
We demonstrate image generation and captioning performance on par with state-of-the-art text-to-image and image-to-text models with orders of magnitude fewer (only 3M) paired image-text data.
no code implementations • 10 Jul 2023 • Cyrus Rashtchian, Charles Herrmann, Chun-Sung Ferng, Ayan Chakrabarti, Dilip Krishnan, Deqing Sun, Da-Cheng Juan, Andrew Tomkins
We find that image-text models (CLIP and ALIGN) are better at recognizing new examples of style transfer than masking-based models (CAN and MAE).
4 code implementations • 1 Jun 2023 • Kihyuk Sohn, Nataniel Ruiz, Kimin Lee, Daniel Castro Chin, Irina Blok, Huiwen Chang, Jarred Barber, Lu Jiang, Glenn Entis, Yuanzhen Li, Yuan Hao, Irfan Essa, Michael Rubinstein, Dilip Krishnan
Pre-trained large text-to-image models synthesize impressive images with an appropriate use of text prompts.
2 code implementations • NeurIPS 2023 • Yonglong Tian, Lijie Fan, Phillip Isola, Huiwen Chang, Dilip Krishnan
We investigate the potential of learning visual representations using synthetic images generated by text-to-image models.
1 code implementation • NeurIPS 2023 • Lijie Fan, Dilip Krishnan, Phillip Isola, Dina Katabi, Yonglong Tian
During training, LaCLIP randomly selects either the original texts or the rewritten versions as text augmentations for each image.
no code implementations • 22 Feb 2023 • Sangnie Bhardwaj, Willie McClinton, Tongzhou Wang, Guillaume Lajoie, Chen Sun, Phillip Isola, Dilip Krishnan
In this paper, we propose a method of learning representations that are instead equivariant to data augmentations.
5 code implementations • 2 Jan 2023 • Huiwen Chang, Han Zhang, Jarred Barber, AJ Maschinot, Jose Lezama, Lu Jiang, Ming-Hsuan Yang, Kevin Murphy, William T. Freeman, Michael Rubinstein, Yuanzhen Li, Dilip Krishnan
Compared to pixel-space diffusion models, such as Imagen and DALL-E 2, Muse is significantly more efficient due to the use of discrete tokens and requiring fewer sampling iterations; compared to autoregressive models, such as Parti, Muse is more efficient due to the use of parallel decoding.
Ranked #1 on Text-to-Image Generation on MS-COCO (FID metric)
1 code implementation • CVPR 2023 • Tianhong Li, Huiwen Chang, Shlok Kumar Mishra, Han Zhang, Dina Katabi, Dilip Krishnan
In this work, we propose MAsked Generative Encoder (MAGE), the first framework to unify SOTA image generation and self-supervised representation learning.
Ranked #3 on Unconditional Image Generation on ImageNet 256x256
1 code implementation • 30 Oct 2022 • Shlok Mishra, Joshua Robinson, Huiwen Chang, David Jacobs, Aaron Sarna, Aaron Maschinot, Dilip Krishnan
Our framework is a minimal and conceptually clean synthesis of (C) contrastive learning, (A) masked autoencoders, and (N) the noise prediction approach used in diffusion models.
1 code implementation • Radiology 2022 • Andrew B. Sellergren, Christina Chen, Zaid Nabulsi, Yuanzhen Li, Aaron Maschinot, Aaron Sarna, Jenny Huang, Charles Lau, Sreenivasa Raju Kalidindi, Mozziyar Etemadi, Florencia Garcia-Vicente, David Melnick, Yun Liu, Krish Eswaran, Daniel Tse, Neeral Beladia, Dilip Krishnan, Shravya Shetty
Supervised contrastive learning enabled performance comparable to state-of-the-art deep learning models in multiple clinical tasks by using as few as 45 images and is a promising method for predictive modeling with use of small data sets and for predicting outcomes in shifting patient populations.
1 code implementation • 1 Dec 2021 • Shlok Mishra, Anshul Shah, Ankan Bansal, Abhyuday Jagannatha, Janit Anjaria, Abhishek Sharma, David Jacobs, Dilip Krishnan
This assumption is mostly satisfied in datasets such as ImageNet where there is a large, centered object, which is highly likely to be present in random crops of the full image.
1 code implementation • CVPR 2022 • Charles Herrmann, Kyle Sargent, Lu Jiang, Ramin Zabih, Huiwen Chang, Ce Liu, Dilip Krishnan, Deqing Sun
In this work, we present pyramid adversarial training (PyramidAT), a simple and effective technique to improve ViT's overall performance.
Ranked #9 on Domain Generalization on ImageNet-C (using extra training data)
no code implementations • 18 Nov 2021 • Apurva Kalia, Dilip Krishnan, Soha Hassoun
Accurately predicting the likelihood of interaction between two objects (compound-protein sequence, user-item, author-paper, etc.)
1 code implementation • 14 Aug 2021 • Andrea Burns, Aaron Sarna, Dilip Krishnan, Aaron Maschinot
Disentangled visual representations have largely been studied with generative models such as Variational AutoEncoders (VAEs).
no code implementations • 15 Mar 2021 • Piotr Teterwak, Chiyuan Zhang, Dilip Krishnan, Michael C. Mozer
We use our reconstruction model as a tool for exploring the nature of representations, including: the influence of model architecture and training objectives (specifically robust losses), the forms of invariance that networks achieve, representational differences between correctly and incorrectly classified images, and the effects of manipulating logits and images.
1 code implementation • NeurIPS 2020 • Yonglong Tian, Chen Sun, Ben Poole, Dilip Krishnan, Cordelia Schmid, Phillip Isola
Contrastive learning between multiple views of the data has recently achieved state of the art performance in the field of self-supervised representation learning.
Ranked #2 on Contrastive Learning on imagenet-1k
24 code implementations • NeurIPS 2020 • Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, Dilip Krishnan
Contrastive learning applied to self-supervised representation learning has seen a resurgence in recent years, leading to state of the art performance in the unsupervised training of deep image models.
Ranked #2 on Class Incremental Learning on cifar100
3 code implementations • ECCV 2020 • Yonglong Tian, Yue Wang, Dilip Krishnan, Joshua B. Tenenbaum, Phillip Isola
The focus of recent meta-learning research has been on the development of learning algorithms that can quickly adapt to test time tasks with limited data and low computational cost.
3 code implementations • ICLR 2020 • Yiding Jiang, Behnam Neyshabur, Hossein Mobahi, Dilip Krishnan, Samy Bengio
We present the first large scale study of generalization in deep networks.
4 code implementations • ICLR 2020 • Yonglong Tian, Dilip Krishnan, Phillip Isola
We demonstrate that this objective ignores important structural knowledge of the teacher network.
Ranked #14 on Knowledge Distillation on CIFAR-100
no code implementations • ICCV 2019 • Piotr Teterwak, Aaron Sarna, Dilip Krishnan, Aaron Maschinot, David Belanger, Ce Liu, William T. Freeman
Image extension models have broad applications in image editing, computational photography and computer graphics.
Ranked #2 on Uncropping on Places2 val
no code implementations • NeurIPS 2019 • Chongli Qin, James Martens, Sven Gowal, Dilip Krishnan, Krishnamurthy Dvijotham, Alhussein Fawzi, Soham De, Robert Stanforth, Pushmeet Kohli
Using this regularizer, we exceed current state of the art and achieve 47% adversarial accuracy for ImageNet with l-infinity adversarial perturbations of radius 4/255 under an untargeted, strong, white-box attack.
8 code implementations • ECCV 2020 • Yonglong Tian, Dilip Krishnan, Phillip Isola
We analyze key properties of the approach that make it work, finding that the contrastive loss outperforms a popular alternative based on cross-view prediction, and that the more views we learn from, the better the resulting representation captures underlying scene semantics.
Ranked #48 on Self-Supervised Action Recognition on UCF101
no code implementations • 10 Jun 2019 • Vighnesh Birodkar, Hossein Mobahi, Dilip Krishnan, Samy Bengio
This operator can learn a strict super-set of what can be learned by average pooling or convolutions.
2 code implementations • ICLR 2019 • Yiding Jiang, Dilip Krishnan, Hossein Mobahi, Samy Bengio
In this paper, we propose such a measure, and conduct extensive empirical studies on how well it can predict the generalization gap.
no code implementations • CVPR 2018 • Tali Dekel, Chuang Gan, Dilip Krishnan, Ce Liu, William T. Freeman
We study the problem of reconstructing an image from information stored at contour locations.
2 code implementations • NeurIPS 2018 • Gamaleldin F. Elsayed, Dilip Krishnan, Hossein Mobahi, Kevin Regan, Samy Bengio
We present a formulation of deep learning that aims at producing a large margin classifier.
no code implementations • 21 Dec 2017 • Tali Dekel, Chuang Gan, Dilip Krishnan, Ce Liu, William T. Freeman
We study the problem of reconstructing an image from information stored at contour locations.
1 code implementation • CVPR 2017 • Forrester Cole, David Belanger, Dilip Krishnan, Aaron Sarna, Inbar Mosseri, William T. Freeman
We present a method for synthesizing a frontal, neutral-expression image of a person's face given an input face photograph.
6 code implementations • CVPR 2017 • Konstantinos Bousmalis, Nathan Silberman, David Dohan, Dumitru Erhan, Dilip Krishnan
Collecting well-annotated image datasets to train modern machine learning algorithms is prohibitively expensive for many tasks.
Generative Adversarial Network Unsupervised Domain Adaptation
5 code implementations • NeurIPS 2016 • Konstantinos Bousmalis, George Trigeorgis, Nathan Silberman, Dilip Krishnan, Dumitru Erhan
However, by focusing only on creating a mapping or shared representation between the two domains, they ignore the individual characteristics of each domain.
Ranked #1 on Domain Adaptation on Synth Objects-to-LINEMOD
no code implementations • ICCV 2015 • Daniel Zoran, Phillip Isola, Dilip Krishnan, William T. Freeman
We demonstrate that this frame- work works well on two important mid-level vision tasks: intrinsic image decomposition and depth from an RGB im- age.
2 code implementations • 21 Nov 2015 • Phillip Isola, Daniel Zoran, Dilip Krishnan, Edward H. Adelson
We propose a self-supervised framework that learns to group visual entities based on their rate of co-occurrence in space and time.
no code implementations • CVPR 2015 • YiChang Shih, Dilip Krishnan, Fredo Durand, William T. Freeman
For single-pane windows, ghosting cues arise from shifted reflections on the two surfaces of the glass pane.
no code implementations • NeurIPS 2014 • Daniel Zoran, Dilip Krishnan, José Bento, Bill Freeman
The Generic Viewpoint Assumption (GVA) states that the position of the viewer or the light in a scene is not special.
no code implementations • 16 Nov 2013 • Dilip Krishnan, Joan Bruna, Rob Fergus
Blind deconvolution has made significant progress in the past decade.
no code implementations • NeurIPS 2009 • Dilip Krishnan, Rob Fergus
In this paper we describe a deconvolution approach that is several orders of magnitude faster than existing techniques that use hyper-Laplacian priors.