no code implementations • ACL (RepL4NLP) 2021 • Pritish Sahu, Michael Cogswell, Ajay Divakaran, Sara Rutherford-Quach
Current pre-trained language models have lots of knowledge, but a more limited ability to use that knowledge.
no code implementations • 2 Jul 2024 • Pritish Sahu, Karan Sikka, Ajay Divakaran
Large Visual Language Models (LVLMs) struggle with hallucinations in visual instruction following task(s), limiting their trustworthiness and real-world applicability.
1 code implementation • 25 Jun 2024 • Yiqiao Jin, Andrew Zhao, Yeon-Chang Lee, Meng Ye, Ajay Divakaran, Srijan Kumar
Our work not only addresses the ongoing challenges in visualizing and analyzing DTDG models but also establishes a foundational framework for future investigations into dynamic graph representation and analysis across various disciplines.
no code implementations • 20 Dec 2023 • Yunye Gong, Robik Shrestha, Jared Claypoole, Michael Cogswell, Arijit Ray, Christopher Kanan, Ajay Divakaran
We propose a novel VQA dataset, BloomVQA, to facilitate comprehensive evaluation of large vision-language models on comprehension tasks.
no code implementations • 30 Nov 2023 • Matthew Gwilliam, Michael Cogswell, Meng Ye, Karan Sikka, Abhinav Shrivastava, Ajay Divakaran
To provide a more thorough evaluation of the capabilities of long video retrieval systems, we propose a pipeline that leverages state-of-the-art large language models to carefully generate a diverse set of synthetic captions for long videos.
no code implementations • CVPR 2024 • Yangyi Chen, Karan Sikka, Michael Cogswell, Heng Ji, Ajay Divakaran
The critique NLF identifies the strengths and weaknesses of the responses and is used to align the LVLMs with human preferences.
1 code implementation • 16 Oct 2023 • Anirudh Som, Karan Sikka, Helen Gent, Ajay Divakaran, Andreas Kathol, Dimitra Vergyri
Paraphrasing of offensive content is a better alternative to content removal and helps improve civility in a communication environment.
no code implementations • 21 Sep 2023 • Yunye Gong, Yi Yao, Xiao Lin, Ajay Divakaran, Melinda Gervasio
Existing conformal prediction algorithms estimate prediction intervals at target confidence levels to characterize the performance of a regression model on new test samples.
1 code implementation • 8 Sep 2023 • Yangyi Chen, Karan Sikka, Michael Cogswell, Heng Ji, Ajay Divakaran
Based on this pipeline and the existing coarse-grained annotated dataset, we build the CURE benchmark to measure both the zero-shot reasoning performance and consistency of VLMs.
no code implementations • ICCV 2023 • Indranil Sur, Karan Sikka, Matthew Walmer, Kaushik Koneripalli, Anirban Roy, Xiao Lin, Ajay Divakaran, Susmit Jha
We present a Multimodal Backdoor Defense technique TIJO (Trigger Inversion using Joint Optimization).
1 code implementation • 7 Apr 2023 • Madeline Schiappa, Raiyaan Abdullah, Shehreen Azad, Jared Claypoole, Michael Cogswell, Ajay Divakaran, Yogesh Rawat
In this work we focus on conceptual understanding of these large V+L models.
1 code implementation • 19 Feb 2023 • Meng Ye, Karan Sikka, Katherine Atwell, Sabit Hassan, Ajay Divakaran, Malihe Alikhani
Content moderation is the process of flagging content based on pre-defined platform rules.
no code implementations • CVPR 2023 • Rohit Gupta, Anirban Roy, Claire Christensen, Sujeong Kim, Sarah Gerard, Madeline Cincebeaux, Ajay Divakaran, Todd Grindal, Mubarak Shah
We learn a class prototype for each class and a loss function is employed to minimize the distances between a class prototype and the samples from the class.
no code implementations • 8 Dec 2022 • Indranil Sur, Zachary Daniels, Abrar Rahman, Kamil Faber, Gianmarco J. Gallardo, Tyler L. Hayes, Cameron E. Taylor, Mustafa Burak Gurbuz, James Smith, Sahana Joshi, Nathalie Japkowicz, Michael Baron, Zsolt Kira, Christopher Kanan, Roberto Corizzo, Ajay Divakaran, Michael Piacentino, Jesse Hostetler, Aswin Raghavan
In this paper, we introduce the Lifelong Reinforcement Learning Components Framework (L2RLCF), which standardizes L2RL systems and assimilates different continual learning components (each addressing different aspects of the lifelong learning problem) into a unified system.
no code implementations • 29 Sep 2022 • Pritish Sahu, Michael Cogswell, Yunye Gong, Ajay Divakaran
The success of Large Language Models (LLMs) indicates they are increasingly able to answer queries like these accurately, but that ability does not necessarily imply a general understanding of concepts relevant to the anchor query.
no code implementations • 9 Aug 2022 • Zachary Daniels, Aswin Raghavan, Jesse Hostetler, Abrar Rahman, Indranil Sur, Michael Piacentino, Ajay Divakaran
We present a version of GR for LRL that satisfies two desiderata: (a) Introspective density modelling of the latent representations of policies learned using deep RL, and (b) Model-free end-to-end learning.
no code implementations • 15 Jul 2022 • Sujeong Kim, Abhinav Garlapati, Jonah Lubin, Amir Tamrakar, Ajay Divakaran
We present a series of two studies conducted to understand user's affective states during voice-based human-machine interactions.
no code implementations • 11 Feb 2022 • Manoj Acharya, Anirban Roy, Kaushik Koneripalli, Susmit Jha, Christopher Kanan, Ajay Divakaran
GCRN consists of two separate graphs to predict object labels based on the contextual cues in the image: 1) a representation graph to learn object features based on the neighboring objects and 2) a context graph to explicitly capture contextual cues from the neighboring objects.
Ranked #1 on Anomaly Detection on COCO-OOC
no code implementations • 22 Oct 2021 • Pritish Sahu, Karan Sikka, Ajay Divakaran
We also observe a drop in performance across all the models when testing on RecipeQA and proposed Meta-RecipeQA (e. g. 83. 6% versus 67. 1% for HTRN), which shows that the proposed dataset is relatively less biased.
no code implementations • 8 Jun 2021 • Pritish Sahu, Michael Cogswell, Sara Rutherford-Quach, Ajay Divakaran
Current pre-trained language models have lots of knowledge, but a more limited ability to use that knowledge.
no code implementations • 20 Apr 2021 • Pritish Sahu, Karan Sikka, Ajay Divakaran
We then evaluate M3C using a textual cloze style question-answering task and highlight an inherent bias in the question answer generation method from [35] that enables a naive baseline to cheat by learning from only answer choices.
1 code implementation • 1 Apr 2021 • Xiao Lin, Meng Ye, Yunye Gong, Giedrius Buracas, Nikoletta Basiou, Ajay Divakaran, Yi Yao
Adapting pre-trained representations has become the go-to recipe for learning new downstream tasks with limited examples.
no code implementations • ICCV 2021 • Yunye Gong, Xiao Lin, Yi Yao, Thomas G. Dietterich, Ajay Divakaran, Melinda Gervasio
Existing calibration algorithms address the problem of covariate shift via unsupervised domain adaptation.
no code implementations • 26 Mar 2021 • Arijit Ray, Michael Cogswell, Xiao Lin, Kamran Alipour, Ajay Divakaran, Yi Yao, Giedrius Burachas
Hence, we propose Error Maps that clarify the error by highlighting image regions where the model is prone to err.
no code implementations • 3 Dec 2020 • Karan Sikka, Indranil Sur, Susmit Jha, Anirban Roy, Ajay Divakaran
We target the problem of detecting Trojans or backdoors in DNNs.
no code implementations • 21 Nov 2020 • Karan Sikka, Jihua Huang, Andrew Silberfarb, Prateeth Nayak, Luke Rohrer, Pritish Sahu, John Byrnes, Ajay Divakaran, Richard Rohwer
We improve zero-shot learning (ZSL) by incorporating common-sense knowledge in DNNs.
no code implementations • 19 Nov 2020 • Meng Ye, Xiao Lin, Giedrius Burachas, Ajay Divakaran, Yi Yao
Few-Shot Learning (FSL) aims to improve a model's generalization capability in low data regimes.
no code implementations • 14 Jul 2020 • Aswin Raghavan, Jesse Hostetler, Indranil Sur, Abrar Rahman, Ajay Divakaran
We propose a wake-sleep cycle of alternating task learning and knowledge consolidation for learning in our framework, and instantiate it for lifelong supervised learning and lifelong RL.
no code implementations • ICML Workshop LifelongML 2020 • Aswin Raghavan, Jesse Hostetler, Indranil Sur, Abrar Rahman, Ajay Divakaran
We propose a wake-sleep cycle of alternating task learning and knowledge consolidation for learning in our framework, and instantiate it for lifelong supervised learning and lifelong RL.
no code implementations • 16 Mar 2020 • Karan Sikka, Andrew Silberfarb, John Byrnes, Indranil Sur, Ed Chow, Ajay Divakaran, Richard Rohwer
We introduce Deep Adaptive Semantic Logic (DASL), a novel framework for automating the generation of deep neural networks that incorporates user-provided formal knowledge to improve learning from data.
no code implementations • ICLR Workshop DeepDiffEq 2019 • Hammad A. Ayyubi, Yi Yao, Ajay Divakaran
Neural Ordinary Differential Equations (NODEs) have proven to be a powerful modeling tool for approximating (interpolation) and forecasting (extrapolation) irregularly sampled time series data.
no code implementations • 25 Sep 2019 • Xiao Lin, Indranil Sur, Samuel A. Nastase, Uri Hasson, Ajay Divakaran, Mohamed R. Amer
Measuring Mutual Information (MI) between high-dimensional, continuous, random variables from observed samples has wide theoretical and practical applications.
no code implementations • IJCNLP 2019 • Arijit Ray, Karan Sikka, Ajay Divakaran, Stefan Lee, Giedrius Burachas
For instance, if a model answers "red" to "What color is the balloon?
1 code implementation • 14 Jul 2019 • Parneet Kaur, Karan Sikka, Weijun Wang, Serge Belongie, Ajay Divakaran
Food classification is a challenging problem due to the large number of categories, high visual similarity between different foods, as well as the lack of datasets for training state-of-the-art deep models.
no code implementations • 17 May 2019 • Karan Sikka, Lucas Van Bramer, Ajay Divakaran
We also show that the user embeddings learned within our joint multimodal embedding model are better at predicting user interests compared to those learned with unimodal content on Instagram data.
no code implementations • 8 May 2019 • Xiao Lin, Indranil Sur, Samuel A. Nastase, Ajay Divakaran, Uri Hasson, Mohamed R. Amer
We demonstrate the effectiveness of our estimators on synthetic benchmarks and a real world fMRI data, with application of inter-subject correlation analysis.
1 code implementation • IJCNLP 2019 • Julia Kruk, Jonah Lubin, Karan Sikka, Xiao Lin, Dan Jurafsky, Ajay Divakaran
Computing author intent from multimodal data like Instagram posts requires modeling a complex relationship between text and image.
no code implementations • 5 Apr 2019 • Arijit Ray, Yi Yao, Rakesh Kumar, Ajay Divakaran, Giedrius Burachas
Our experiments, therefore, demonstrate that ExAG is an effective means to evaluate the efficacy of AI-generated explanations on a human-AI collaborative task.
no code implementations • ICCV 2019 • Samyak Datta, Karan Sikka, Anirban Roy, Karuna Ahuja, Devi Parikh, Ajay Divakaran
We propose a novel end-to-end model that uses caption-to-image retrieval as a `downstream' task to guide the process of phrase localization.
no code implementations • 26 Nov 2018 • Pallabi Ghosh, Yi Yao, Larry S. Davis, Ajay Divakaran
We show results on CAD120 (which provides pre-computed node features and edge weights for fair performance comparison across algorithms) as well as a more complex real-world activity dataset, Charades.
no code implementations • 4 Jul 2018 • Karuna Ahuja, Karan Sikka, Anirban Roy, Ajay Divakaran
We show that our model outperforms other baselines on the benchmark Ad dataset and also show qualitative results to highlight the advantages of using multihop co-attention.
no code implementations • ECCV 2018 • Ankan Bansal, Karan Sikka, Gaurav Sharma, Rama Chellappa, Ajay Divakaran
We introduce and tackle the problem of zero-shot object detection (ZSD), which aims to detect object classes which are not observed during training.
no code implementations • 23 Dec 2017 • Parneet Kaur, Karan Sikka, Ajay Divakaran
Food classification from images is a fine-grained classification problem.
no code implementations • 6 May 2015 • Mohamed R. Amer, Behjat Siddiquie, Amir Tamrakar, David A. Salter, Brian Lande, Darius Mehri, Ajay Divakaran
We present a novel approach to computational modeling of social interactions based on modeling of essential social interaction predicates (ESIPs) such as joint attention and entrainment.