Distilling Knowledge from Reader to Retriever for Question Answering

ICLR 2021  ·  Gautier Izacard, Edouard Grave ·

The task of information retrieval is an important component of many natural language processing systems, such as open domain question answering. While traditional methods were based on hand-crafted features, continuous representations based on neural networks recently obtained competitive results. A challenge of using such methods is to obtain supervised data to train the retriever model, corresponding to pairs of query and support documents. In this paper, we propose a technique to learn retriever models for downstream tasks, inspired by knowledge distillation, and which does not require annotated pairs of query and documents. Our approach leverages attention scores of a reader model, used to solve the task based on retrieved documents, to obtain synthetic labels for the retriever. We evaluate our method on question answering, obtaining state-of-the-art results.

PDF Abstract ICLR 2021 PDF ICLR 2021 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Question Answering NarrativeQA FiD+Distil BLEU-1 35.3 # 7
BLEU-4 7.5 # 8
METEOR 11.1 # 8
Rouge-L 32 # 9
Question Answering TriviaQA FiD+Distil EM 72.1 # 19

Methods


No methods listed for this paper. Add relevant methods here