Browse > Reasoning > Natural Language Visual Grounding

Natural Language Visual Grounding

4 papers with code · Reasoning

Leaderboards

No evaluation results yet. Help compare methods by submit evaluation metrics.

Latest papers with code

Learning Cross-modal Context Graph for Visual Grounding

AAAI-2020 2020 youngfly11/LCMCG-PyTorch

To address their limitations, this paper proposes a language-guided graph representation to capture the global context of grounding entities and their relations, and develop a cross-modal graph matching strategy for the multiple-phrase visual grounding task.

GRAPH MATCHING LANGUAGE MODELLING NATURAL LANGUAGE VISUAL GROUNDING PHRASE GROUNDING

8
13 Feb 2020

Searching for Ambiguous Objects in Videos using Relational Referring Expressions

3 Aug 2019hazananayurt/viref

Especially in ambiguous settings, humans prefer expressions (called relational referring expressions) that describe an object with respect to a distinguishing, unique object.

DEEP ATTENTION NATURAL LANGUAGE VISUAL GROUNDING

5
03 Aug 2019

Self-Monitoring Navigation Agent via Auxiliary Progress Estimation

ICLR 2019 chihyaoma/selfmonitoring-agent

The Vision-and-Language Navigation (VLN) task entails an agent following navigational instruction in photo-realistic unknown environments.

NATURAL LANGUAGE VISUAL GROUNDING VISION-LANGUAGE NAVIGATION VISUAL NAVIGATION

90
10 Jan 2019

Grounding of Textual Phrases in Images by Reconstruction

12 Nov 2015akirafukui/vqa-mcb

We propose a novel approach which learns grounding by reconstructing a given phrase using an attention mechanism, which can be either latent or optimized directly.

LANGUAGE MODELLING NATURAL LANGUAGE VISUAL GROUNDING

174
12 Nov 2015