no code implementations • 23 Feb 2023 • Yang Chen, Hexiang Hu, Yi Luan, Haitian Sun, Soravit Changpinyo, Alan Ritter, Ming-Wei Chang
Our analysis shows that it is challenging for the state-of-the-art multi-modal pre-trained models to answer visual information seeking questions, but this capability is improved through fine-tuning on the automated InfoSeek dataset.
no code implementations • 22 Feb 2023 • Hexiang Hu, Yi Luan, Yang Chen, Urvashi Khandelwal, Mandar Joshi, Kenton Lee, Kristina Toutanova, Ming-Wei Chang
Large-scale multi-modal pre-training models such as CLIP and PaLI exhibit strong generalization on various visual domains and tasks.
no code implementations • 23 Sep 2022 • Zhuyun Dai, Vincent Y. Zhao, Ji Ma, Yi Luan, Jianmo Ni, Jing Lu, Anton Bakalov, Kelvin Guu, Keith B. Hall, Ming-Wei Chang
To amplify the power of a few examples, we propose Prompt-base Query Generation for Retriever (Promptagator), which leverages large language models (LLM) as a few-shot query generator, and creates task-specific retrievers based on the generated data.
no code implementations • 12 Apr 2022 • Ivan Stelmakh, Yi Luan, Bhuwan Dhingra, Ming-Wei Chang
In contrast to existing long-form QA tasks (such as ELI5), ASQA admits a clear notion of correctness: a user faced with a good summary should be able to answer different interpretations of the original ambiguous question.
no code implementations • 16 Dec 2021 • Zeqiu Wu, Yi Luan, Hannah Rashkin, David Reitter, Hannaneh Hajishirzi, Mari Ostendorf, Gaurav Singh Tomar
Compared to standard retrieval tasks, passage retrieval for conversational question answering (CQA) poses new challenges in understanding the current user question, as each question needs to be interpreted within the dialogue context.
2 code implementations • 15 Dec 2021 • Jianmo Ni, Chen Qu, Jing Lu, Zhuyun Dai, Gustavo Hernández Ábrego, Ji Ma, Vincent Y. Zhao, Yi Luan, Keith B. Hall, Ming-Wei Chang, Yinfei Yang
With multi-stage training, surprisingly, scaling up the model size brings significant improvement on a variety of retrieval tasks, especially for out-of-domain generalization.
Ranked #9 on
Zero-shot Text Search
on BEIR
1 code implementation • 1 May 2020 • Yi Luan, Jacob Eisenstein, Kristina Toutanova, Michael Collins
Dual encoders perform retrieval by encoding documents and queries into dense lowdimensional vectors, scoring each document by its inner product with the query.
no code implementations • 24 Apr 2020 • Mandar Joshi, Kenton Lee, Yi Luan, Kristina Toutanova
We present a method to represent input texts by contextualizing them jointly with dynamically retrieved textual encyclopedic background knowledge from multiple documents.
2 code implementations • IJCNLP 2019 • David Wadden, Ulme Wennberg, Yi Luan, Hannaneh Hajishirzi
We examine the capabilities of a unified, multi-task framework for three information extraction tasks: named entity recognition, relation extraction, and event extraction.
Ranked #5 on
Joint Entity and Relation Extraction
on SciERC
2 code implementations • ACL 2019 • Qingyun Wang, Lifu Huang, Zhiying Jiang, Kevin Knight, Heng Ji, Mohit Bansal, Yi Luan
We present a PaperRobot who performs as an automatic research assistant by (1) conducting deep understanding of a large collection of human-written papers in a target domain and constructing comprehensive background knowledge graphs (KGs); (2) creating new ideas by predicting links from the background KGs, by combining graph attention and contextual text attention; (3) incrementally writing some key elements of a new paper based on memory-attention networks: from the input title along with predicted related entities to generate a paper abstract, from the abstract to generate conclusion and future work, and finally from future work to generate a title for a follow-on paper.
3 code implementations • NAACL 2019 • Yi Luan, Dave Wadden, Luheng He, Amy Shah, Mari Ostendorf, Hannaneh Hajishirzi
We introduce a general framework for several information extraction tasks that share span representations using dynamically constructed span graphs.
Ranked #1 on
Relation Extraction
on ACE 2004
(Cross Sentence metric)
Joint Entity and Relation Extraction
Named Entity Recognition (NER)
3 code implementations • NAACL 2019 • Rik Koncel-Kedziorski, Dhanush Bekal, Yi Luan, Mirella Lapata, Hannaneh Hajishirzi
Generating texts which express complex ideas spanning multiple sentences requires a structured representation of their content (document plan), but these representations are prohibitively expensive to manually produce.
Ranked #6 on
KG-to-Text Generation
on AGENDA
no code implementations • 19 Sep 2018 • Yonghui Huang, Yunhui Li, Yi Luan
This work improves monolingual sentence alignment for text simplification, specifically for text in standard and simple Wikipedia.
3 code implementations • EMNLP 2018 • Yi Luan, Luheng He, Mari Ostendorf, Hannaneh Hajishirzi
We introduce a multi-task setup of identifying and classifying entities, relations, and coreference clusters in scientific articles.
Ranked #7 on
Joint Entity and Relation Extraction
on SciERC
Coreference Resolution
Joint Entity and Relation Extraction
+1
no code implementations • 26 Aug 2018 • Yi Luan, Mari Ostendorf, Hannaneh Hajishirzi
This paper describes our submission for the SemEval 2018 Task 7 shared task on semantic relation extraction and classification in scientific papers.
no code implementations • SEMEVAL 2018 • Yi Luan, Mari Ostendorf, Hannaneh Hajishirzi
This paper describes our submission for SemEval 2018 Task 7 shared task on semantic relation extraction and classification in scientific papers.
no code implementations • IJCNLP 2017 • Yi Luan, Chris Brockett, Bill Dolan, Jianfeng Gao, Michel Galley
Building a persona-based conversation agent is challenging owing to the lack of large amounts of speaker-specific conversation data for model training.
no code implementations • EMNLP 2017 • Yi Luan, Mari Ostendorf, Hannaneh Hajishirzi
This paper addresses the problem of extracting keyphrases from scientific articles and categorizing them as corresponding to a task, process, or material.
1 code implementation • 31 Mar 2016 • Yi Luan, Yangfeng Ji, Mari Ostendorf
In this paper, we present a conversational model that incorporates both context and participant role for two-party conversations.