Iterative Alternating Neural Attention for Machine Reading

7 Jun 2016  ·  Alessandro Sordoni, Philip Bachman, Adam Trischler, Yoshua Bengio ·

We propose a novel neural attention architecture to tackle machine comprehension tasks, such as answering Cloze-style queries with respect to a document. Unlike previous models, we do not collapse the query into a single vector, instead we deploy an iterative alternating attention mechanism that allows a fine-grained exploration of both the query and the document. Our model outperforms state-of-the-art baselines in standard machine comprehension benchmarks such as CNN news articles and the Children's Book Test (CBT) dataset.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Question Answering Children's Book Test AIA Accuracy-NE 72% # 3
Question Answering CNN / Daily Mail AIA CNN 76.1 # 5

Methods


No methods listed for this paper. Add relevant methods here