no code implementations • 24 Jan 2024 • Ke Ye, Heinrich Jiang, Afshin Rostamizadeh, Ayan Chakrabarti, Giulia Desalvo, Jean-François Kagy, Lazaros Karydas, Gui Citovsky, Sanjiv Kumar
In this paper, we present SpacTor, a new training procedure consisting of (1) a hybrid objective combining span corruption (SC) and token replacement detection (RTD), and (2) a two-stage curriculum that optimizes the hybrid objective over the initial $\tau$ iterations, then transitions to standard SC loss.
3 code implementations • CVPR 2024 • Sadeep Jayasumana, Srikumar Ramalingam, Andreas Veit, Daniel Glasner, Ayan Chakrabarti, Sanjiv Kumar
It is an unbiased estimator that does not make any assumptions on the probability distribution of the embeddings and is sample efficient.
no code implementations • CVPR 2024 • Sadeep Jayasumana, Daniel Glasner, Srikumar Ramalingam, Andreas Veit, Ayan Chakrabarti, Sanjiv Kumar
Modern text-to-image generation models produce high-quality images that are both photorealistic and faithful to the text prompts.
no code implementations • 10 Jul 2023 • Cyrus Rashtchian, Charles Herrmann, Chun-Sung Ferng, Ayan Chakrabarti, Dilip Krishnan, Deqing Sun, Da-Cheng Juan, Andrew Tomkins
We find that image-text models (CLIP and ALIGN) are better at recognizing new examples of style transfer than masking-based models (CAN and MAE).
1 code implementation • NeurIPS 2023 • Florian Stimberg, Ayan Chakrabarti, Chun-Ta Lu, Hussein Hazimeh, Otilia Stretcu, Wei Qiao, Yintao Liu, Merve Kaya, Cyrus Rashtchian, Ariel Fuxman, Mehmet Tek, Sven Gowal
We evaluate 33 pretrained models on the benchmark and train models with different augmentations, architectures and training methods on subsets of the obfuscations to measure generalization.
1 code implementation • 31 Jul 2022 • Jiaming Qiu, Ruiqi Wang, Ayan Chakrabarti, Roch Guerin, Chenyang Lu
Because of limited computing capacity, embedded devices rely on a parsimonious classification model with uneven accuracy.
1 code implementation • 21 Oct 2021 • Mingyang Xie, Manav Kulshrestha, Shaojie Wang, Jinghan Yang, Ayan Chakrabarti, Ning Zhang, Yevgeniy Vorobeychik
Modern AI tools, such as generative adversarial networks, have transformed our ability to create and modify visual data with photorealistic results.
1 code implementation • 13 Oct 2021 • Srinadh Bhojanapalli, Ayan Chakrabarti, Andreas Veit, Michal Lukasik, Himanshu Jain, Frederick Liu, Yin-Wen Chang, Sanjiv Kumar
Pairwise dot product-based attention allows Transformers to exchange information between tokens in an input-dependent way, and is key to their success across diverse applications in language and vision.
no code implementations • 16 Jun 2021 • Srinadh Bhojanapalli, Ayan Chakrabarti, Himanshu Jain, Sanjiv Kumar, Michal Lukasik, Andreas Veit
State-of-the-art transformer models use pairwise dot-product based self-attention, which comes at a computational cost quadratic in the input sequence length.
no code implementations • ICCV 2021 • Srinadh Bhojanapalli, Ayan Chakrabarti, Daniel Glasner, Daliang Li, Thomas Unterthiner, Andreas Veit
We find that when pre-trained with a sufficient amount of data, ViT models are at least as robust as the ResNet counterparts on a broad range of perturbations.
no code implementations • CVPR 2021 • Zhihao Xia, Michaël Gharbi, Federico Perazzi, Kalyan Sunkavalli, Ayan Chakrabarti
We introduce a neural network-based method to denoise pairs of images taken in quick succession, with and without a flash, in low-light environments.
1 code implementation • 26 Oct 2020 • Ayan Chakrabarti, Roch Guérin, Chenyang Lu, Jiangnan Liu
To deploy machine learning-based algorithms for real-time applications with strict latency constraints, we consider an edge-computing setting where a subset of inputs are offloaded to the edge for processing by an accurate but resource-intensive model, and the rest are processed only by a less-accurate model on the device itself.
1 code implementation • 17 Oct 2020 • Jinghan Yang, Adith Boloor, Ayan Chakrabarti, Xuan Zhang, Yevgeniy Vorobeychik
We propose a scalable approach for finding adversarial modifications of a simulated autonomous driving environment using a differentiable approximation for the mapping from environmental modifications (rectangles on the road) to the corresponding video inputs to the controller neural network.
no code implementations • 23 Jun 2020 • Shaojie Wang, Tong Wu, Ayan Chakrabarti, Yevgeniy Vorobeychik
First, we find that the fusion model is usually both more accurate, and more robust against single-source attacks than single-sensor deep neural networks.
no code implementations • 21 Mar 2020 • Francesco Pittaluga, Zaid Tasneem, Justin Folden, Brevin Tilmon, Ayan Chakrabarti, Sanjeev J. Koppal
We present a proof-of-concept LIDAR design that allows adaptive real-time measurements according to dynamically specified measurement patterns.
no code implementations • CVPR 2020 • Zhihao Xia, Federico Perazzi, Michaël Gharbi, Kalyan Sunkavalli, Ayan Chakrabarti
Bursts of images exhibit significant self-similarity across both time and space.
1 code implementation • 4 Dec 2019 • Jinghan Yang, Ayan Chakrabarti, Yevgeniy Vorobeychik
We study this problem formally as a combinatorial optimization problem in the context of geolocation prediction facilitated by deep learning.
1 code implementation • NeurIPS 2019 • Zhihao Xia, Ayan Chakrabarti
Deep neural networks have been very successful in compressive-sensing and image restoration applications, as a means to estimate images from partial, blurry, or otherwise degraded measurements.
no code implementations • 28 Nov 2019 • Weidong Cao, Liu Ke, Ayan Chakrabarti, Xuan Zhang
Recent works propose neural network- (NN-) inspired analog-to-digital converters (NNADCs) and demonstrate their great potentials in many emerging applications.
1 code implementation • 13 Jun 2019 • Zhihao Xia, Ayan Chakrabarti
We evaluate our method for training networks for compressive-sensing and blind deconvolution, considering both non-blind and blind training for the latter.
1 code implementation • CVPR 2020 • Zhihao Xia, Patrick Sullivan, Ayan Chakrabarti
Beyond depth estimation from a single image, the monocular cue is useful in a broader range of depth inference applications and settings---such as when one can leverage other available depth cues for improved accuracy.
Ranked #64 on Monocular Depth Estimation on NYU-Depth V2 (RMSE metric)
1 code implementation • 8 Mar 2019 • Kyle Yee, Ayan Chakrabarti
Modern neural network-based algorithms are able to produce highly accurate depth estimates from stereo image pairs, nearly matching the reliability of measurements from more expensive depth sensors.
1 code implementation • ICLR 2019 • Ayan Chakrabarti, Benjamin Moseley
Training convolutional neural network models is memory intensive since back-propagation requires storing activations of all intermediate layers.
no code implementations • CVPR 2019 • Zhuo Hui, Ayan Chakrabarti, Kalyan Sunkavalli, Aswin C. Sankaranarayanan
We present a method to separate a single image captured under two illuminants, with different spectra, into the two images corresponding to the appearance of the scene under each individual illuminant.
1 code implementation • 13 Jun 2018 • Zhihao Xia, Ayan Chakrabarti
In this work, we propose a new method for natural image denoising that trains a deep neural network to determine whether patches in a noisy image input share common underlying patterns.
no code implementations • 14 Feb 2018 • Francesco Pittaluga, Sanjeev J. Koppal, Ayan Chakrabarti
We present a framework to learn privacy-preserving encodings of images that inhibit inference of chosen private attributes, while allowing recovery of other desirable information.
3 code implementations • ICLR 2018 • Charles Schaff, David Yunis, Ayan Chakrabarti, Matthew R. Walter
The physical design of a robot and the policy that controls its motion are inherently coupled, and should be determined according to the task and environment.
2 code implementations • ICLR 2018 • Behnam Neyshabur, Srinadh Bhojanapalli, Ayan Chakrabarti
Training generative adversarial networks is unstable in high-dimensions as the true data distribution tends to be concentrated in a small fraction of the ambient space.
1 code implementation • 24 Mar 2017 • Charles Schaff, David Yunis, Ayan Chakrabarti, Matthew R. Walter
The accuracy of such a beacon-based localization system depends both on how beacons are distributed in the environment, and how the robot's location is inferred based on noisy and potentially ambiguous measurements.
no code implementations • 17 Nov 2016 • Igor Vasiljevic, Ayan Chakrabarti, Gregory Shakhnarovich
We investigate the extent to which this degradation is due to the mismatch between training and input image statistics.
no code implementations • 14 Sep 2016 • Ayan Chakrabarti, Kalyan Sunkavalli
We present a single-shot system to recover surface geometry of objects with spatially-varying albedos, from images captured under a calibrated RGB photometric stereo setup---with three light directions multiplexed across different color channels in the observed RGB image.
no code implementations • NeurIPS 2016 • Ayan Chakrabarti, Jingyu Shao, Gregory Shakhnarovich
A single color image can contain many cues informative towards different aspects of local geometric structure.
no code implementations • NeurIPS 2016 • Ayan Chakrabarti
In this paper, we specifically consider the design and inference problems in a typical color camera---where the sensor is able to measure only one color channel at each pixel location, and computational inference is required to reconstruct a full color image.
2 code implementations • 15 Mar 2016 • Ayan Chakrabarti
We present a new method for blind motion deblurring that uses a neural network trained to compute estimates of sharp image patches from observations that are blurred by an unknown motion kernel.
no code implementations • NeurIPS 2015 • Ayan Chakrabarti
Specifically, we describe an illuminant estimation method that is built around a "classifier" for identifying the true chromaticity of a pixel given its luminance (absolute brightness across color channels).
no code implementations • CVPR 2015 • Ayan Chakrabarti, Ying Xiong, Steven J. Gortler, Todd Zickler
We introduce a multi-scale framework for low-level vision, where the goal is estimating physical scene values from image data---such as depth from stereo image pairs.
no code implementations • 27 Nov 2013 • Ayan Chakrabarti, Ying Xiong, Baochen Sun, Trevor Darrell, Daniel Scharstein, Todd Zickler, Kate Saenko
To produce images that are suitable for display, tone-mapping is widely used in digital cameras to map linear color measurements into narrow gamuts with limited dynamic range.
no code implementations • 10 Oct 2013 • Ying Xiong, Ayan Chakrabarti, Ronen Basri, Steven J. Gortler, David W. Jacobs, Todd Zickler
We develop a framework for extracting a concise representation of the shape information available from diffuse shading in a small image patch.