Search Results for author: Arjun R. Akula

Found 8 papers, 3 papers with code

Question Generation for Evaluating Cross-Dataset Shifts in Multi-modal Grounding

no code implementations24 Jan 2022 Arjun R. Akula

Visual question answering (VQA) is the multi-modal task of answering natural language questions about an input image.

Question Answering Question Generation +2

Words aren't enough, their order matters: On the Robustness of Grounding Visual Referring Expressions

1 code implementation ACL 2020 Arjun R. Akula, Spandana Gella, Yaser Al-Onaizan, Song-Chun Zhu, Siva Reddy

To measure the true progress of existing models, we split the test set into two sets, one which requires reasoning on linguistic structure and the other which doesn't.

Contrastive Learning Multi-Task Learning +2

Discourse Parsing in Videos: A Multi-modal Appraoch

1 code implementation6 Mar 2019 Arjun R. Akula, Song-Chun Zhu

We propose the task of Visual Discourse Parsing, which requires understanding discourse relations among scenes in a video.

Discourse Parsing Visual Dialog +1

Cannot find the paper you are looking for? You can Submit a new open access paper.