Detecting Visual Relationships with Deep Relational Networks

CVPR 2017  ·  Bo Dai, Yuqi Zhang, Dahua Lin ·

Relationships among objects play a crucial role in image understanding. Despite the great success of deep learning techniques in recognizing individual objects, reasoning about the relationships among objects remains a challenging task. Previous methods often treat this as a classification problem, considering each type of relationship (e.g. "ride") or each distinct visual phrase (e.g. "person-ride-horse") as a category. Such approaches are faced with significant difficulties caused by the high diversity of visual appearance for each kind of relationships or the large number of distinct visual phrases. We propose an integrated framework to tackle this problem. At the heart of this framework is the Deep Relational Network, a novel formulation designed specifically for exploiting the statistical dependencies between objects and their relationships. On two large datasets, the proposed method achieves substantial improvement over state-of-the-art.

PDF Abstract CVPR 2017 PDF CVPR 2017 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Visual Relationship Detection VRD Phrase Detection Dai et. al [[Dai, Zhang, and Lin2017]] R@100 23.45 # 3
R@50 19.93 # 4
Visual Relationship Detection VRD Predicate Detection Dai et. al [[Dai, Zhang, and Lin2017]] R@100 81.90 # 4
R@50 80.78 # 4
Visual Relationship Detection VRD Relationship Detection Dai et. al [[Dai, Zhang, and Lin2017]] R@100 20.88 # 4
R@50 17.73 # 5

Methods


No methods listed for this paper. Add relevant methods here