Syntax-based Attention Model for Natural Language Inference

Introducing attentional mechanism in neural network is a powerful concept, and has achieved impressive results in many natural language processing tasks. However, most of the existing models impose attentional distribution on a flat topology, namely the entire input representation sequence... (read more)

Results in Papers With Code
(↓ scroll down to see all results)