Grounded Situation Recognition with Transformers

19 Nov 2021  ·  Junhyeong Cho, Youngseok Yoon, Hyeonjun Lee, Suha Kwak ·

Grounded Situation Recognition (GSR) is the task that not only classifies a salient action (verb), but also predicts entities (nouns) associated with semantic roles and their locations in the given image. Inspired by the remarkable success of Transformers in vision tasks, we propose a GSR model based on a Transformer encoder-decoder architecture. The attention mechanism of our model enables accurate verb classification by capturing high-level semantic feature of an image effectively, and allows the model to flexibly deal with the complicated and image-dependent relations between entities for improved noun classification and localization. Our model is the first Transformer architecture for GSR, and achieves the state of the art in every evaluation metric on the SWiG benchmark. Our code is available at https://github.com/jhcho99/gsrtr .

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Situation Recognition imSitu GSRTR Top-1 Verb 40.63 # 5
Top-1 Verb & Value 32.15 # 4
Top-5 Verbs 69.81 # 4
Top-5 Verbs & Value 54.13 # 5
Grounded Situation Recognition SWiG GSRTR Top-1 Verb 40.63 # 5
Top-1 Verb & Value 32.15 # 5
Top-1 Verb & Grounded-Value 25.49 # 4
Top-5 Verbs 69.81 # 4
Top-5 Verbs & Value 54.13 # 5
Top-5 Verbs & Grounded-Value 42.5 # 4

Methods