no code implementations • 15 Sep 2024 • Qilong Zhangli, Di Liu, Abhishek Aich, Dimitris Metaxas, Samuel Schulter
Notably, on four benchmark datasets with label space inconsistencies during inference, we outperform previous methods by 1. 6% mIoU for semantic segmentation, 9. 1% PQ for panoptic segmentation, 12. 1% AP for instance segmentation, and 3. 0% in the newly proposed PIQ metric.
no code implementations • 23 Apr 2024 • Abhishek Aich, Yumin Suh, Samuel Schulter, Manmohan Chandraker
With efficiency being a high priority for scaling such models, we observed that the state-of-the-art method Mask2Former uses ~50% of its compute only on the transformer encoder.
no code implementations • 23 Apr 2024 • Manyi Yao, Abhishek Aich, Yumin Suh, Amit Roy-Chowdhury, Christian Shelton, Manmohan Chandraker
The third step is to use the aforementioned derived dataset to train a gating network that predicts the number of encoder layers to be used, conditioned on the input image.
no code implementations • ICCV 2023 • Abhishek Aich, Samuel Schulter, Amit K. Roy-Chowdhury, Manmohan Chandraker, Yumin Suh
Further, we present a simple but effective search algorithm that translates user constraints to runtime width configurations of both the shared encoder and task decoders, for sampling the sub-architectures.
no code implementations • 14 Dec 2022 • Abhishek Aich, Kuan-Chuan Peng, Amit K. Roy-Chowdhury
Most cross-domain unsupervised Video Anomaly Detection (VAD) works assume that at least few task-relevant target domain training data are available for adaptation from the source to the target domain.
no code implementations • 20 Sep 2022 • Abhishek Aich, Shasha Li, Chengyu Song, M. Salman Asif, Srikanth V. Krishnamurthy, Amit K. Roy-Chowdhury
Our goal is to design an attack strategy that can learn from such natural scenes by leveraging the local patch differences that occur inherently in such images (e. g. difference between the local patch on the object `person' and the object `bike' in a traffic scene).
no code implementations • 20 Sep 2022 • Abhishek Aich, Calvin-Khang Ta, Akash Gupta, Chengyu Song, Srikanth V. Krishnamurthy, M. Salman Asif, Amit K. Roy-Chowdhury
Using the joint image-text features to train the generator, we show that GAMA can craft potent transferable perturbations in order to fool victim classifiers in various attack settings.
1 code implementation • 4 Jun 2022 • Calvin-Khang Ta, Abhishek Aich, Akash Gupta, Amit K. Roy-Chowdhury
In this work, we explore a sparsity and dictionary learning-based approach and present a novel self-supervised learning method for single-image denoising where the noise is approximated as a Poisson process, requiring no clean ground-truth data.
1 code implementation • NeurIPS 2021 • Shasha Li, Abhishek Aich, Shitong Zhu, M. Salman Asif, Chengyu Song, Amit K. Roy-Chowdhury, Srikanth V. Krishnamurthy
When compared to the image classification models, black-box adversarial attacks against video classification models have been largely understudied.
1 code implementation • 29 Jul 2021 • Akash Gupta, Abhishek Aich, Kevin Rodriguez, G. Venugopala Reddy, Amit K. Roy-Chowdhury
In this paper, we propose a data-driven Deep Quantized Latent Representation (DQLR) methodology for high-quality image reconstruction in the Shoot Apical Meristem (SAM) of Arabidopsis thaliana.
no code implementations • ICCV 2021 • Abhishek Aich, Meng Zheng, Srikrishna Karanam, Terrence Chen, Amit K. Roy-Chowdhury, Ziyan Wu
To alleviate these problems, we propose Spatio-Temporal Representation Factorization (STRF), a flexible new computational unit that can be used in conjunction with most existing 3D convolutional neural network architectures for re-ID.
Ranked #2 on Person Re-Identification on DukeMTMC-VideoReID
no code implementations • 10 May 2021 • Abhishek Aich
In this report, we present a theoretical support of the continual learning method \textbf{Elastic Weight Consolidation}, introduced in paper titled `Overcoming catastrophic forgetting in neural networks'.
no code implementations • 31 Aug 2020 • Akash Gupta, Abhishek Aich, Amit K. Roy-Chowdhury
Different from these works, we address a more realistic problem of high frame-rate sharp video synthesis with no prior assumption that input is always blurry.
1 code implementation • CVPR 2020 • Abhishek Aich, Akash Gupta, Rameswar Panda, Rakib Hyder, M. Salman Asif, Amit K. Roy-Chowdhury
Different from these methods, we focus on the problem of generating videos from latent noise vectors, without any reference input frames.