Several Experiments on Investigating Pretraining and Knowledge-Enhanced Models for Natural Language Inference

27 Apr 2019 Tianda Li Xiaodan Zhu Quan Liu Qian Chen Zhigang Chen Si Wei

Natural language inference (NLI) is among the most challenging tasks in natural language understanding. Recent work on unsupervised pretraining that leverages unsupervised signals such as language-model and sentence prediction objectives has shown to be very effective on a wide range of NLP problems... (read more)

PDF Abstract

Results from the Paper

  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods used in the Paper

🤖 No Methods Found Help the community by adding them if they're not listed; e.g. Deep Residual Learning for Image Recognition uses ResNet