no code implementations • ICML 2020 • Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat, Ming-Wei Chang
Language model pre-training has been shown to capture a surprising amount of world knowledge, crucial for NLP tasks such as question answering.
no code implementations • EMNLP 2020 • Adam Fisch, Kenton Lee, Ming-Wei Chang, Jonathan Clark, Regina Barzilay
In this task, we use question-answer (QA) pairs{---}a natural expression of information need{---}from users, instead of reference captions, for both training and post-inference evaluation.
no code implementations • 22 Feb 2023 • Hexiang Hu, Yi Luan, Yang Chen, Urvashi Khandelwal, Mandar Joshi, Kenton Lee, Kristina Toutanova, Ming-Wei Chang
Large-scale multi-modal pre-training models such as CLIP and PaLI exhibit strong generalization on various visual domains and tasks.
no code implementations • 20 Dec 2022 • Fangyu Liu, Julian Martin Eisenschlos, Francesco Piccinno, Syrine Krichene, Chenxi Pang, Kenton Lee, Mandar Joshi, Wenhu Chen, Nigel Collier, Yasemin Altun
Compared with a SOTA model finetuned on more than >28k data points, DePlot+LLM with just one-shot prompting achieves a 24. 0% improvement over finetuned SOTA on human-written queries from the task of chart QA.
Ranked #1 on
Chart Question Answering
on ChartQA
no code implementations • 19 Dec 2022 • Fangyu Liu, Francesco Piccinno, Syrine Krichene, Chenxi Pang, Kenton Lee, Mandar Joshi, Yasemin Altun, Nigel Collier, Julian Martin Eisenschlos
Visual language data such as plots, charts, and infographics are ubiquitous in the human world.
Ranked #1 on
Visual Question Answering (VQA)
on PlotQA-D1
no code implementations • 7 Oct 2022 • Kenton Lee, Mandar Joshi, Iulia Turc, Hexiang Hu, Fangyu Liu, Julian Eisenschlos, Urvashi Khandelwal, Peter Shaw, Ming-Wei Chang, Kristina Toutanova
Visually-situated language is ubiquitous -- sources range from textbooks with diagrams to web pages with images and tables, to mobile apps with buttons and forms.
Ranked #3 on
Chart Question Answering
on ChartQA
1 code implementation • 2 Apr 2022 • David Wadden, Nikita Gupta, Kenton Lee, Kristina Toutanova
We introduce the task of entity-centric query refinement.
no code implementations • 30 Jun 2021 • Iulia Turc, Kenton Lee, Jacob Eisenstein, Ming-Wei Chang, Kristina Toutanova
Zero-shot cross-lingual transfer is emerging as a practical solution: pre-trained models later fine-tuned on one transfer language exhibit surprising performance when tested on many target languages.
no code implementations • EMNLP 2021 • Sewon Min, Kenton Lee, Ming-Wei Chang, Kristina Toutanova, Hannaneh Hajishirzi
We study multi-answer retrieval, an under-explored problem that requires retrieving passages to cover multiple distinct answers for a given question.
1 code implementation • 2 Feb 2021 • Kenton Lee, Kelvin Guu, Luheng He, Tim Dozat, Hyung Won Chung
In many applications of machine learning, certain categories of examples may be underrepresented in the training data, causing systems to underperform on such "few-shot" cases at test time.
no code implementations • 1 Jan 2021 • Sewon Min, Jordan Boyd-Graber, Chris Alberti, Danqi Chen, Eunsol Choi, Michael Collins, Kelvin Guu, Hannaneh Hajishirzi, Kenton Lee, Jennimaria Palomaki, Colin Raffel, Adam Roberts, Tom Kwiatkowski, Patrick Lewis, Yuxiang Wu, Heinrich Küttler, Linqing Liu, Pasquale Minervini, Pontus Stenetorp, Sebastian Riedel, Sohee Yang, Minjoon Seo, Gautier Izacard, Fabio Petroni, Lucas Hosseini, Nicola De Cao, Edouard Grave, Ikuya Yamada, Sonse Shimaoka, Masatoshi Suzuki, Shumpei Miyawaki, Shun Sato, Ryo Takahashi, Jun Suzuki, Martin Fajcik, Martin Docekal, Karel Ondrej, Pavel Smrz, Hao Cheng, Yelong Shen, Xiaodong Liu, Pengcheng He, Weizhu Chen, Jianfeng Gao, Barlas Oguz, Xilun Chen, Vladimir Karpukhin, Stan Peshterliev, Dmytro Okhonko, Michael Schlichtkrull, Sonal Gupta, Yashar Mehdad, Wen-tau Yih
We review the EfficientQA competition from NeurIPS 2020.
1 code implementation • 9 Nov 2020 • Adam Fisch, Kenton Lee, Ming-Wei Chang, Jonathan H. Clark, Regina Barzilay
In this task, we use question-answer (QA) pairs---a natural expression of information need---from users, instead of reference captions, for both training and post-inference evaluation.
no code implementations • EMNLP 2020 • Gabriel Ilharco, Cesar Ilharco, Iulia Turc, Tim Dettmers, Felipe Ferreira, Kenton Lee
Scale has played a central role in the rapid progress natural language processing has enjoyed in recent years.
3 code implementations • NAACL 2021 • Akari Asai, Jungo Kasai, Jonathan H. Clark, Kenton Lee, Eunsol Choi, Hannaneh Hajishirzi
Multilingual question answering tasks typically assume answers exist in the same language as the question.
no code implementations • ACL 2020 • Alane Suhr, Ming-Wei Chang, Peter Shaw, Kenton Lee
We study the task of cross-database semantic parsing (XSP), where a system that maps natural language utterances to executable SQL queries is evaluated on databases unseen during training.
1 code implementation • ACL 2020 • Hao Cheng, Ming-Wei Chang, Kenton Lee, Kristina Toutanova
We address the problem of extractive question answering using document-level distant super-vision, pairing questions and relevant documents with answer strings.
no code implementations • 24 Apr 2020 • Mandar Joshi, Kenton Lee, Yi Luan, Kristina Toutanova
We present a method to represent input texts by contextualizing them jointly with dynamically retrieved textual encyclopedic background knowledge from multiple documents.
6 code implementations • 10 Feb 2020 • Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat, Ming-Wei Chang
Language model pre-training has been shown to capture a surprising amount of world knowledge, crucial for NLP tasks such as question answering.
Ranked #8 on
Question Answering
on WebQuestions
no code implementations • IJCNLP 2019 • Daniel Andor, Luheng He, Kenton Lee, Emily Pitler
Reading comprehension models have been successfully applied to extractive text answers, but it is unclear how best to generalize these models to abstractive numerical answers.
Ranked #3 on
Question Answering
on DROP Test
40 code implementations • ICLR 2020 • Iulia Turc, Ming-Wei Chang, Kenton Lee, Kristina Toutanova
Recent developments in natural language representations have been accompanied by large and expensive models that leverage vast amounts of general-domain text through self-supervised pre-training.
3 code implementations • ACL 2019 • Lajanugen Logeswaran, Ming-Wei Chang, Kenton Lee, Kristina Toutanova, Jacob Devlin, Honglak Lee
First, we show that strong reading comprehension models pre-trained on large unlabeled data can be used to generalize to unseen entities.
3 code implementations • ACL 2019 • Kenton Lee, Ming-Wei Chang, Kristina Toutanova
We show for the first time that it is possible to jointly learn the retriever and reader from question-answer string pairs and without any IR system.
Ranked #9 on
Question Answering
on WebQuestions
1 code implementation • Transactions of the Association of Computational Linguistics 2019 • Tom Kwiatkowski, Jennimaria Palomaki, Olivia Redfield, Michael Collins, Ankur Parikh, Chris Alberti, Danielle Epstein, Illia Polosukhin, Jacob Devlin, Kenton Lee, Kristina Toutanova, Llion Jones, Matthew Kelcey, Ming-Wei Chang, Andrew M. Dai, Jakob Uszkoreit, Quoc Le, Slav Petrov
The public release consists of 307, 373 training examples with single annotations, 7, 830 examples with 5-way annotations for development data, and a further 7, 842 examples 5-way annotated sequestered as test data.
Ranked #7 on
Question Answering
on Natural Questions (long)
1 code implementation • NAACL 2019 • Christopher Clark, Kenton Lee, Ming-Wei Chang, Tom Kwiatkowski, Michael Collins, Kristina Toutanova
In this paper we study yes/no questions that are naturally occurring --- meaning that they are generated in unprompted and unconstrained settings.
no code implementations • ICLR 2019 • Ming-Wei Chang, Kristina Toutanova, Kenton Lee, Jacob Devlin
Hierarchical neural architectures are often used to capture long-distance dependencies and have been applied to many document-level tasks such as summarization, document segmentation, and sentiment analysis.
3 code implementations • 24 Jan 2019 • Chris Alberti, Kenton Lee, Michael Collins
This technical note describes a new baseline for the Natural Questions.
Ranked #6 on
Question Answering
on Natural Questions (long)
no code implementations • 5 Nov 2018 • Hao Cheng, Ming-Wei Chang, Kenton Lee, Ankur Parikh, Michael Collins, Kristina Toutanova
We study approaches to improve fine-grained short answer Question Answering models by integrating coarse-grained data annotated for paragraph-level relevance and show that coarsely annotated data can bring significant performance gains.
505 code implementations • NAACL 2019 • Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova
We introduce a new language representation model called BERT, which stands for Bidirectional Encoder Representations from Transformers.
Ranked #1 on
Question Answering
on CoQA
1 code implementation • EMNLP 2018 • Swabha Swayamdipta, Sam Thomson, Kenton Lee, Luke Zettlemoyer, Chris Dyer, Noah A. Smith
We introduce the syntactic scaffold, an approach to incorporating syntactic information into semantic tasks.
1 code implementation • ACL 2018 • Luheng He, Kenton Lee, Omer Levy, Luke Zettlemoyer
Recent BIO-tagging-based neural semantic role labeling models are very high performing, but assume gold predicates as part of the input and cannot incorporate span-level features.
no code implementations • ACL 2018 • Omer Levy, Kenton Lee, Nicholas FitzGerald, Luke Zettlemoyer
LSTMs were introduced to combat vanishing gradients in simple RNNs by augmenting them with gated additive recurrent connections.
5 code implementations • NAACL 2018 • Kenton Lee, Luheng He, Luke Zettlemoyer
We introduce a fully differentiable approximation to higher-order inference for coreference resolution.
Ranked #14 on
Coreference Resolution
on OntoNotes
44 code implementations • NAACL 2018 • Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, Luke Zettlemoyer
We introduce a new type of deep contextualized word representation that models both (1) complex characteristics of word use (e. g., syntax and semantics), and (2) how these uses vary across linguistic contexts (i. e., to model polysemy).
Ranked #3 on
Citation Intent Classification
on ACL-ARC
(using extra training data)
Citation Intent Classification
Conversational Response Selection
+7
4 code implementations • EMNLP 2017 • Kenton Lee, Luheng He, Mike Lewis, Luke Zettlemoyer
We introduce the first end-to-end coreference resolution model and show that it significantly outperforms all previous work without using a syntactic parser or hand-engineered mention detector.
Ranked #15 on
Coreference Resolution
on CoNLL 2012
1 code implementation • ACL 2017 • Luheng He, Kenton Lee, Mike Lewis, Luke Zettlemoyer
We introduce a new deep learning model for semantic role labeling (SRL) that significantly improves the state of the art, along with detailed analyses to reveal its strengths and limitations.
Ranked #2 on
Predicate Detection
on CoNLL 2005
2 code implementations • 21 May 2017 • Kenton Lee, Omer Levy, Luke Zettlemoyer
We introduce recurrent additive networks (RANs), a new gated RNN which is distinguished by the use of purely additive latent state updates.
1 code implementation • 4 Nov 2016 • Kenton Lee, Shimi Salant, Tom Kwiatkowski, Ankur Parikh, Dipanjan Das, Jonathan Berant
In this paper, we focus on this answer extraction task, presenting a novel model architecture that efficiently builds fixed length representations of all spans in the evidence document with a recurrent network.
Ranked #43 on
Question Answering
on SQuAD1.1 dev
1 code implementation • EMNLP 2016 • Kenton Lee, Mike Lewis, Luke Zettlemoyer
We introduce the first global recursive neural parsing model with optimality guarantees during decoding.