no code implementations • 12 Mar 2024 • Pranav Singh Chib, Pravendra Singh
We introduce LG-Traj, a novel approach incorporating LLMs to generate motion cues present in pedestrian past/observed trajectories.
no code implementations • 27 Nov 2023 • Nikhil Kumar, Pravendra Singh
This is the first review in the field of small and dim target detection in infrared imagery, encompassing various methodologies ranging from conventional image processing to cutting-edge deep learning-based approaches.
no code implementations • 10 Oct 2023 • Suruchi Kumari, Pravendra Singh
This paper conducts a thorough review of data-efficient deep learning methods for medical image analysis.
no code implementations • 29 Sep 2023 • Pranav Singh Chib, Pravendra Singh
Motion prediction systems must effectively learn spatial and temporal information from the past to forecast the future trajectories of the agent.
no code implementations • 18 Jul 2023 • Suruchi Kumari, Pravendra Singh
In this paper, we provide a comprehensive literature review of recent deep UDA approaches in medical imaging from a technical perspective.
no code implementations • 10 Jul 2023 • Md Junaid Mahmood, Pranaw Raj, Divyansh Agarwal, Suruchi Kumari, Pravendra Singh
To evaluate the performance of our proposed approach, we conduct experiments on two publicly available medical image classification benchmark datasets: the skin lesion classification (ISIC 2018) and the blood cell classification dataset (BCCD).
2 code implementations • 10 Jul 2023 • Pranav Singh Chib, Pravendra Singh
End-to-End driving is a promising paradigm as it circumvents the drawbacks associated with modular systems, such as their overwhelming complexity and propensity for error propagation.
no code implementations • 19 Oct 2022 • Pravendra Singh, Pratik Mazumder, Mohammed Asad Karim
However, in the future, some classes may become restricted due to privacy/ethical concerns, and the restricted class knowledge has to be removed from the models that have been trained on them.
no code implementations • 23 Dec 2021 • Mohammed Asad Karim, Indu Joshi, Pratik Mazumder, Pravendra Singh
We apply our proposed approach to state-of-the-art class-incremental learning methods and empirically show that our framework significantly improves the performance of these methods.
no code implementations • 29 Sep 2021 • Pratik Mazumder, Pravendra Singh, Mohammed Asad Karim
A naive solution is to simply train the model from scratch on the complete training data while leaving out the training samples from the restricted classes (FDR - full data retraining).
no code implementations • 30 Jun 2021 • Pratik Mazumder, Pravendra Singh, Vinay P. Namboodiri
Our approach significantly improves their performance and further reduces the model biases in the limited data regime.
no code implementations • 12 Jun 2021 • Mohammed Asad Karim, Vinay Kumar Verma, Pravendra Singh, Vinay Namboodiri, Piyush Rai
In our approach, we learn robust representations that are generalizable across tasks without suffering from the problems of catastrophic forgetting and overfitting to accommodate future classes with limited samples.
no code implementations • CVPR 2021 • Pravendra Singh, Pratik Mazumder, Piyush Rai, Vinay P. Namboodiri
Our proposed method uses weight rectifications and affine transformations in order to adapt the model to different tasks that arrive sequentially.
no code implementations • 1 Mar 2021 • Pratik Mazumder, Pravendra Singh, Piyush Rai
Our method selects very few parameters from the model for training every new set of classes instead of training the full model.
no code implementations • NeurIPS 2020 • Pravendra Singh, Vinay Kumar Verma, Pratik Mazumder, Lawrence Carin, Piyush Rai
Further, our approach does not require storing data samples from the old tasks, which is done by many replay based methods.
no code implementations • 22 Nov 2020 • Pratik Mazumder, Pravendra Singh, Vinay P. Namboodiri
Our method relies on generating robust prototypes from a set of few examples.
no code implementations • 29 Jun 2020 • Pratik Mazumder, Pravendra Singh, Vinay P. Namboodiri
We then simultaneously train for the composite rotation prediction task along with the original classification task, which forces the network to learn high-quality generic features that help improve the few-shot classification performance.
no code implementations • 8 Jun 2020 • Pravendra Singh, Pratik Mazumder, Vinay P. Namboodiri
Our proposed technique, namely Passive Batch Injection Training Technique (PBITT), even reduces the level of overfitting in networks that already use the standard techniques for reducing overfitting such as $L_2$ regularization and batch normalization, resulting in significant accuracy improvements.
no code implementations • 27 May 2020 • Pratik Mazumder, Pravendra Singh, Kranti Kumar Parida, Vinay P. Namboodiri
We use the semantic relatedness of text embeddings as a means for zero-shot learning by aligning audio and video embeddings with the corresponding class label text feature space.
Ranked #6 on GZSL Video Classification on ActivityNet-GZSL(main)
no code implementations • 26 May 2020 • Rajat, Munender Varshney, Pravendra Singh, Vinay P. Namboodi
The approach we adopt is one of active learning, i. e., incrementally selecting a set of samples that need supervision based on the current model, obtaining supervision for these samples, retraining the model with the additional set of supervised samples and proceeding again to select the next set of samples.
no code implementations • 15 Jan 2020 • Vinay Kumar Verma, Pravendra Singh, Vinay P. Namboodiri, Piyush Rai
The pruner is essentially a multitask deep neural network with binary outputs that help identify the filters from each layer of the original network that do not have any significant contribution to the model and can therefore be pruned.
no code implementations • 5 Jan 2020 • Pravendra Singh, Munender Varshney, Vinay P. Namboodiri
In this paper, we have proposed a cooperative initialization for training the deep network using ReLU activation function to improve the network performance.
no code implementations • 21 Oct 2019 • Pratik Mazumder, Pravendra Singh, Vinay Namboodiri
We propose an alternative design for pointwise convolution, which uses spatial information from the input efficiently.
1 code implementation • 11 May 2019 • Pravendra Singh, Vinay Kumar Verma, Piyush Rai, Vinay P. Namboodiri
Our framework, called Play and Prune (PP), jointly prunes and fine-tunes CNN model parameters, with an adaptive pruning rate, while maintaining the model's predictive performance.
no code implementations • 11 Mar 2019 • Pravendra Singh, Pratik Mazumder, Vinay P. Namboodiri
Recently researchers have tried to boost the performance of CNNs by re-calibrating the feature maps produced by these filters, e. g., Squeeze-and-Excitation Networks (SENets).
1 code implementation • CVPR 2019 • Pravendra Singh, Vinay Kumar Verma, Piyush Rai, Vinay P. Namboodiri
We present a novel deep learning architecture in which the convolution operation leverages heterogeneous kernels.
no code implementations • 26 Nov 2018 • Pravendra Singh, Vinay Kumar Verma, Piyush Rai, Vinay P. Namboodiri
We present a filter correlation based model compression approach for deep convolutional neural networks.
no code implementations • 20 Nov 2018 • Pravendra Singh, Vinay Sameer Raja Kadi, Nikhil Verma, Vinay P. Namboodiri
Convolutional neural networks (CNN) have achieved impressive performance on the wide variety of tasks (classification, detection, etc.)
no code implementations • 20 Nov 2018 • Pravendra Singh, Manikandan. R, Neeraj Matiyali, Vinay P. Namboodiri
Additionally, we also empirically show our method's adaptability for classification based architecture VGG16 on datasets CIFAR and German Traffic Sign Recognition Benchmark (GTSRB) achieving a compression rate of 125X and 200X with the reduction in flops by 90. 50% and 96. 6% respectively with no loss of accuracy.
1 code implementation • 28 Mar 2015 • Pravendra Singh
An intelligent version of the sliding-puzzle game is developed using the new Go programming language, which uses a concurrent version of the A* Informed Search Algorithm to power solver-bot that runs in the background.