no code implementations • EMNLP 2017 • Daniel Hewlett, Llion Jones, Alex Lacoste, re, Izzeddin Gur
We also evaluate the model in a semi-supervised setting by downsampling the WikiReading training set to create increasingly smaller amounts of supervision, while leaving the full unlabeled document corpus to train a sequence autoencoder on document windows.
no code implementations • ACL 2017 • Eunsol Choi, Daniel Hewlett, Jakob Uszkoreit, Illia Polosukhin, Alex Lacoste, re, Jonathan Berant
We present a framework for question answering that can efficiently scale to longer documents while maintaining or even improving performance of state-of-the-art models.