Search Results for author: Arjun R. Akula

Found 8 papers, 3 papers with code

Question Generation for Evaluating Cross-Dataset Shifts in Multi-modal Grounding

no code implementations24 Jan 2022 Arjun R. Akula

Visual question answering (VQA) is the multi-modal task of answering natural language questions about an input image.

Question Answering Question Generation +2

CX-ToM: Counterfactual Explanations with Theory-of-Mind for Enhancing Human Trust in Image Recognition Models

1 code implementation3 Sep 2021 Arjun R. Akula, Keze Wang, Changsong Liu, Sari Saba-Sadiya, Hongjing Lu, Sinisa Todorovic, Joyce Chai, Song-Chun Zhu

More concretely, our CX-ToM framework generates sequence of explanations in a dialog by mediating the differences between the minds of machine and human user.

Explainable Artificial Intelligence (XAI)

Words aren't enough, their order matters: On the Robustness of Grounding Visual Referring Expressions

1 code implementation ACL 2020 Arjun R. Akula, Spandana Gella, Yaser Al-Onaizan, Song-Chun Zhu, Siva Reddy

To measure the true progress of existing models, we split the test set into two sets, one which requires reasoning on linguistic structure and the other which doesn't.

Contrastive Learning Multi-Task Learning +2

Discourse Parsing in Videos: A Multi-modal Appraoch

1 code implementation6 Mar 2019 Arjun R. Akula, Song-Chun Zhu

We propose the task of Visual Discourse Parsing, which requires understanding discourse relations among scenes in a video.

Discourse Parsing Visual Dialog +1

Cannot find the paper you are looking for? You can Submit a new open access paper.