no code implementations • 6 Mar 2024 • Fangyuan Xu, Kyle Lo, Luca Soldaini, Bailey Kuehl, Eunsol Choi, David Wadden
To evaluate the capabilities of current LLMs on this task, we construct KIWI, a dataset of knowledge-intensive writing instructions in the scientific domain.
no code implementations • 18 Oct 2023 • Hung-Ting Chen, Fangyuan Xu, Shane A. Arora, Eunsol Choi
Our study provides new insights on how retrieval augmentation impacts long, knowledge-rich text generation of LMs.
1 code implementation • 6 Oct 2023 • Fangyuan Xu, Weijia Shi, Eunsol Choi
Retrieving documents and prepending them in-context at inference time improves performance of language model (LMs) on a wide range of tasks.
1 code implementation • 30 May 2023 • Abhilash Potluri, Fangyuan Xu, Eunsol Choi
Long-form question answering systems provide rich information by presenting paragraph-level answers, often containing optional background or auxiliary information.
1 code implementation • 29 May 2023 • Fangyuan Xu, Yixiao Song, Mohit Iyyer, Eunsol Choi
We present a careful analysis of experts' evaluation, which focuses on new aspects such as the comprehensiveness of the answer.
no code implementations • NAACL 2022 • Shufan Wang, Fangyuan Xu, Laure Thompson, Eunsol Choi, Mohit Iyyer
We show that not only do state-of-the-art LFQA models struggle to generate relevant examples, but also that standard evaluation metrics such as ROUGE are insufficient to judge exemplification quality.
1 code implementation • ACL 2022 • Fangyuan Xu, Junyi Jessy Li, Eunsol Choi
Long-form answers, consisting of multiple sentences, can provide nuanced and comprehensive answers to a broader set of questions.