Densely Connected Attention Propagation for Reading Comprehension

NeurIPS 2018  ·  Yi Tay, Luu Anh Tuan, Siu Cheung Hui, Jian Su ·

We propose DecaProp (Densely Connected Attention Propagation), a new densely connected neural architecture for reading comprehension (RC). There are two distinct characteristics of our model. Firstly, our model densely connects all pairwise layers of the network, modeling relationships between passage and query across all hierarchical levels. Secondly, the dense connectors in our network are learned via attention instead of standard residual skip-connectors. To this end, we propose novel Bidirectional Attention Connectors (BAC) for efficiently forging connections throughout the network. We conduct extensive experiments on four challenging RC benchmarks. Our proposed approach achieves state-of-the-art results on all four, outperforming existing baselines by up to $2.6\%-14.2\%$ in absolute F1 score.

PDF Abstract NeurIPS 2018 PDF NeurIPS 2018 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Question Answering NarrativeQA DecaProp BLEU-1 44.35 # 3
BLEU-4 27.61 # 2
METEOR 21.80 # 3
Rouge-L 44.69 # 5
Question Answering NewsQA DecaProp F1 66.3 # 4
EM 53.1 # 2
Open-Domain Question Answering Quasar DecaProp EM (Quasar-T) 38.6 # 3
F1 (Quasar-T) 46.9 # 3
Question Answering Quasart-T DECAPROP EM 38.6 # 6
Open-Domain Question Answering SearchQA DecaProp Unigram Acc 62.2 # 1
N-gram F1 70.8 # 1
EM 56.8 # 7
F1 63.6 # 3
Open-Domain Question Answering SearchQA DECAPROP EM 62.2 # 5

Methods


No methods listed for this paper. Add relevant methods here