Case-based Reasoning for Natural Language Queries over Knowledge Bases

It is often challenging to solve a complex problem from scratch, but much easier if we can access other similar problems with their solutions -- a paradigm known as case-based reasoning (CBR). We propose a neuro-symbolic CBR approach (CBR-KBQA) for question answering over large knowledge bases. CBR-KBQA consists of a nonparametric memory that stores cases (question and logical forms) and a parametric model that can generate a logical form for a new question by retrieving cases that are relevant to it. On several KBQA datasets that contain complex questions, CBR-KBQA achieves competitive performance. For example, on the ComplexWebQuestions dataset, CBR-KBQA outperforms the current state of the art by 11\% on accuracy. Furthermore, we show that CBR-KBQA is capable of using new cases \emph{without} any further training: by incorporating a few human-labeled examples in the case memory, CBR-KBQA is able to successfully generate logical forms containing unseen KB entities as well as relations.

PDF Abstract EMNLP 2021 PDF EMNLP 2021 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Knowledge Base Question Answering ComplexWebQuestions PullNet Accuracy 45.9 # 4
Knowledge Base Question Answering ComplexWebQuestions QGG Accuracy 44.1 # 5
Knowledge Base Question Answering ComplexWebQuestions CBR-KBQA Accuracy 70.4 # 2
Semantic Parsing WebQuestionsSP CBR-KBQA Accuracy 70 # 3


No methods listed for this paper. Add relevant methods here