ScanQA: 3D Question Answering for Spatial Scene Understanding

We propose a new 3D spatial understanding task of 3D Question Answering (3D-QA). In the 3D-QA task, models receive visual information from the entire 3D scene of the rich RGB-D indoor scan and answer the given textual questions about the 3D scene. Unlike the 2D-question answering of VQA, the conventional 2D-QA models suffer from problems with spatial understanding of object alignment and directions and fail the object identification from the textual questions in 3D-QA. We propose a baseline model for 3D-QA, named ScanQA model, where the model learns a fused descriptor from 3D object proposals and encoded sentence embeddings. This learned descriptor correlates the language expressions with the underlying geometric features of the 3D scan and facilitates the regression of 3D bounding boxes to determine described objects in textual questions and outputs correct answers. We collected human-edited question-answer pairs with free-form answers that are grounded to 3D objects in each 3D scene. Our new ScanQA dataset contains over 40K question-answer pairs from the 800 indoor scenes drawn from the ScanNet dataset. To the best of our knowledge, the proposed 3D-QA task is the first large-scale effort to perform object-grounded question-answering in 3D environments.

PDF Abstract CVPR 2022 PDF CVPR 2022 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Benchmark
3D Question Answering (3D-QA) ScanQA Test w/ objects ScanQA Exact Match 23.45 # 3
BLEU-1 31.56 # 6
BLEU-4 12.04 # 3
ROUGE 34.34 # 6
METEOR 13.55 # 5
CIDEr 67.29 # 4
3D Question Answering (3D-QA) ScanQA Test w/ objects VoteNet+MCAN Exact Match 19.71 # 6
BLEU-1 29.46 # 7
BLEU-4 6.08 # 8
ROUGE 30.97 # 7
METEOR 12.07 # 7
CIDEr 58.23 # 7
3D Question Answering (3D-QA) ScanQA Test w/ objects ScanRefer+MCAN Exact Match 20.56 # 5
BLEU-1 27.85 # 8
BLEU-4 7.46 # 7
ROUGE 30.68 # 8
METEOR 11.97 # 8
CIDEr 57.56 # 8

Methods


No methods listed for this paper. Add relevant methods here