SQA3D: Situated Question Answering in 3D Scenes

14 Oct 2022  ·  Xiaojian Ma, Silong Yong, Zilong Zheng, Qing Li, Yitao Liang, Song-Chun Zhu, Siyuan Huang ·

We propose a new task to benchmark scene understanding of embodied agents: Situated Question Answering in 3D Scenes (SQA3D). Given a scene context (e.g., 3D scan), SQA3D requires the tested agent to first understand its situation (position, orientation, etc.) in the 3D scene as described by text, then reason about its surrounding environment and answer a question under that situation. Based upon 650 scenes from ScanNet, we provide a dataset centered around 6.8k unique situations, along with 20.4k descriptions and 33.4k diverse reasoning questions for these situations. These questions examine a wide spectrum of reasoning capabilities for an intelligent agent, ranging from spatial relation comprehension to commonsense understanding, navigation, and multi-hop reasoning. SQA3D imposes a significant challenge to current multi-modal especially 3D reasoning models. We evaluate various state-of-the-art approaches and find that the best one only achieves an overall score of 47.20%, while amateur human participants can reach 90.06%. We believe SQA3D could facilitate future embodied AI research with stronger situation understanding and reasoning capability.

PDF Abstract

Datasets


Introduced in the Paper:

SQA3D

Used in the Paper:

ScanNet

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Question Answering SQA3D ScanQA AnswerExactMatch (Question Answering) 46.58 # 4
Question Answering SQA3D ScanQA (w/ auxiliary loss) AnswerExactMatch (Question Answering) 47.20 # 3
Referring Expression SQA3D Random Acc@0.5m 14.60 # 1
Acc@1.0m 34.21 # 1
Acc@15° 22.39 # 1
Acc@30° 42.28 # 1

Methods


No methods listed for this paper. Add relevant methods here