Learning Natural Language Inference using Bidirectional LSTM model and Inner-Attention

30 May 2016  ·  Yang Liu, Chengjie Sun, Lei Lin, Xiaolong Wang ·

In this paper, we proposed a sentence encoding-based model for recognizing text entailment. In our approach, the encoding of sentence is a two-stage process. Firstly, average pooling was used over word-level bidirectional LSTM (biLSTM) to generate a first-stage sentence representation. Secondly, attention mechanism was employed to replace average pooling on the same sentence for better representations. Instead of using target sentence to attend words in source sentence, we utilized the sentence's first-stage representation to attend words appeared in itself, which is called "Inner-Attention" in our paper . Experiments conducted on Stanford Natural Language Inference (SNLI) Corpus has proved the effectiveness of "Inner-Attention" mechanism. With less number of parameters, our model outperformed the existing best sentence encoding-based approach by a large margin.

PDF Abstract

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Natural Language Inference SNLI 600D (300+300) BiLSTM encoders with intra-attention and symbolic preproc. % Test Accuracy 85.0 # 74
% Train Accuracy 85.9 # 64
Parameters 2.8m # 4
Natural Language Inference SNLI 600D (300+300) BiLSTM encoders with intra-attention % Test Accuracy 84.2 # 82
% Train Accuracy 84.5 # 69
Parameters 2.8m # 4
Natural Language Inference SNLI 600D (300+300) BiLSTM encoders % Test Accuracy 83.3 # 86
% Train Accuracy 86.4 # 62
Parameters 2.0m # 4

Methods