2 code implementations • ACL 2022 • Saadia Gabriel, Skyler Hallinan, Maarten Sap, Pemi Nguyen, Franziska Roesner, Eunsol Choi, Yejin Choi
Even to a simple and short news headline, readers react in a multitude of ways: cognitively (e. g. inferring the writer’s intent), emotionally (e. g. feeling distrust), and behaviorally (e. g. sharing the news with their friends).
1 code implementation • Findings (EMNLP) 2021 • Jifan Chen, Eunsol Choi, Greg Durrett
To build robust question answering systems, we need the ability to verify whether answers to questions are truly correct, not just “good enough” in the context of imperfect QA datasets.
no code implementations • NAACL (DADC) 2022 • Venelin Kovatchev, Trina Chatterjee, Venkata S Govindarajan, Jifan Chen, Eunsol Choi, Gabriella Chronis, Anubrata Das, Katrin Erk, Matthew Lease, Junyi Jessy Li, Yating Wu, Kyle Mahowald
Developing methods to adversarially challenge NLP systems is a promising avenue for improving both model performance and interpretability.
no code implementations • 3 Dec 2024 • James Allan, Eunsol Choi, Daniel P. Lopresti, Hamed Zamani
In the fast-evolving field of information retrieval (IR), the integration of generative AI technologies such as large language models (LLMs) is transforming how users search for and interact with information.
no code implementations • 8 Nov 2024 • Fangyuan Xu, Tanya Goyal, Eunsol Choi
Generating long sequences of tokens given a long-context input imposes a heavy computational burden for large language models (LLMs).
1 code implementation • 26 Oct 2024 • Atula Tejaswi, Yoonsang Lee, Sujay Sanghavi, Eunsol Choi
Our approach, RARe, finetunes a pre-trained model with in-context examples whose query is semantically similar to the target query.
no code implementations • 18 Oct 2024 • Michael JQ Zhang, Zhilin Wang, Jena D. Hwang, Yi Dong, Olivier Delalleau, Yejin Choi, Eunsol Choi, Xiang Ren, Valentina Pyatkin
We find that the majority of disagreements are in opposition with standard reward modeling approaches, which are designed with the assumption that annotator disagreement is noise.
no code implementations • 17 Oct 2024 • Michael J. Q. Zhang, W. Bradley Knox, Eunsol Choi
This allows LLMs to learn to ask clarifying questions when it can generate responses that are tailored to each user interpretation in future turns.
no code implementations • 7 Oct 2024 • Aniruddh Sriram, Fangyuan Xu, Eunsol Choi, Greg Durrett
By leveraging the AVeriTeC dataset, which annotates subquestions for claims with human written answers from evidence documents, we fine-tune Contriever with a contrastive objective based on multiple training signals, including distillation from GPT-4, evaluating subquestion answers, and gold labels in the dataset.
no code implementations • 26 Sep 2024 • Hung-Ting Chen, Eunsol Choi
On this data, retrievers paired with a corpus are evaluated to surface a document set that contains diverse perspectives.
no code implementations • 12 Aug 2024 • Mina Huh, Fangyuan Xu, Yi-Hao Peng, Chongyan Chen, Hansika Murugu, Danna Gurari, Eunsol Choi, Amy Pavel
Vision language models can now generate long-form answers to questions about images - long-form visual question answers (LFVQA).
no code implementations • 8 Jul 2024 • Zeyu Leo Liu, Shrey Pandit, Xi Ye, Eunsol Choi, Greg Durrett
An instance in our benchmark consists of a synthetic API function update paired with a program synthesis example that uses the updated functionality; our goal is to update an LLM to be able to solve this program synthesis example without providing documentation of the update at inference time.
1 code implementation • 25 Jun 2024 • Shane Arora, Marzena Karpinska, Hung-Ting Chen, Ipsita Bhattacharjee, Mohit Iyyer, Eunsol Choi
To bridge this gap, we introduce CaLMQA, a collection of 1. 5K complex culturally specific questions spanning 23 languages and 51 culturally agnostic questions translated from English into 22 other languages.
1 code implementation • 25 Jun 2024 • Thom Lake, Eunsol Choi, Greg Durrett
Alignment suppresses irrelevant and unhelpful content while shifting the output distribution toward longer responses that cover information spanning several responses from the base LLM, essentially presenting diverse information in a single response.
1 code implementation • 20 Jun 2024 • Atula Tejaswi, Nilesh Gupta, Eunsol Choi
In this paper, we study building language-specific LLMs by adapting monolingual and multilingual LLMs.
1 code implementation • 30 May 2024 • Vijay Lingam, Atula Tejaswi, Aditya Vavre, Aneesh Shetty, Gautham Krishna Gudur, Joydeep Ghosh, Alex Dimakis, Eunsol Choi, Aleksandar Bojchevski, Sujay Sanghavi
Extensive experiments on language and vision benchmarks show that SVFT recovers up to 96% of full fine-tuning performance while training only 0. 006 to 0. 25% of parameters, outperforming existing methods that only recover up to 85% performance using 0. 03 to 0. 8% of the trainable parameter budget.
no code implementations • 18 Apr 2024 • Yoonsang Lee, Xi Ye, Eunsol Choi
and a set of documents discussing different people named Michael Jordan, can LMs distinguish entity mentions to generate a cohesive answer to the question?
no code implementations • 6 Mar 2024 • Fangyuan Xu, Kyle Lo, Luca Soldaini, Bailey Kuehl, Eunsol Choi, David Wadden
To evaluate the capabilities of current LLMs on this task, we construct KIWI, a dataset of knowledge-intensive writing instructions in the scientific domain.
no code implementations • 2 Feb 2024 • Zhisheng Zheng, Puyuan Peng, Ziyang Ma, Xie Chen, Eunsol Choi, David Harwath
By integrating Spatial-AST with LLaMA-2 7B model, BAT transcends standard Sound Event Localization and Detection (SELD) tasks, enabling the model to reason about the relationships between the sounds in its environment.
no code implementations • 16 Nov 2023 • Michael J. Q. Zhang, Eunsol Choi
In this work, we study such behavior in LMs by proposing a task-agnostic framework for resolving ambiguity by asking users clarifying questions.
1 code implementation • 16 Nov 2023 • Yoonsang Lee, Pranav Atreya, Xi Ye, Eunsol Choi
We perform analysis on three multi-answer question answering datasets, which allows us to further study answer set ordering strategies based on the LM's knowledge of each answer.
no code implementations • 18 Oct 2023 • Hung-Ting Chen, Fangyuan Xu, Shane A. Arora, Eunsol Choi
Our study provides new insights on how retrieval augmentation impacts long, knowledge-rich text generation of LMs.
2 code implementations • 6 Oct 2023 • Fangyuan Xu, Weijia Shi, Eunsol Choi
Retrieving documents and prepending them in-context at inference time improves performance of language model (LMs) on a wide range of tasks.
no code implementations • 14 Aug 2023 • Xuewen Yao, Miriam Mikhelson, S. Craig Watkins, Eunsol Choi, Edison Thomaz, Kaya de Barbaro
In collaboration with Postpartum Support International (PSI), a non-profit organization dedicated to supporting caregivers with postpartum mood and anxiety disorders, we developed three chatbots to provide context-specific empathetic support to postpartum caregivers, leveraging both rule-based and generative models.
1 code implementation • NeurIPS 2023 • Shankar Padmanabhan, Yasumasa Onoe, Michael J. Q. Zhang, Greg Durrett, Eunsol Choi
Then, we update the model parameters so that the distribution of the LM (the student) matches the distribution of the LM conditioned on the definition (the teacher) on the transfer set.
1 code implementation • 14 Jun 2023 • Anuj Diwan, Eunsol Choi, David Harwath
We present the first unified study of the efficiency of self-attention-based Transformer variants spanning text, speech and vision.
1 code implementation • 30 May 2023 • Abhilash Potluri, Fangyuan Xu, Eunsol Choi
Long-form question answering systems provide rich information by presenting paragraph-level answers, often containing optional background or auxiliary information.
1 code implementation • 29 May 2023 • Fangyuan Xu, Yixiao Song, Mohit Iyyer, Eunsol Choi
We present a careful analysis of experts' evaluation, which focuses on new aspects such as the comprehensiveness of the answer.
2 code implementations • 24 May 2023 • Anuj Diwan, Anirudh Srinivasan, David Harwath, Eunsol Choi
We first pretrain a model on large-scale monolingual speech data, finetune it with a small amount of parallel speech data (20-60 hours), and lastly train with an unsupervised backtranslation objective.
1 code implementation • 24 May 2023 • Michael J. Q. Zhang, Eunsol Choi
While large language models are able to retain vast amounts of world knowledge seen during pretraining, such knowledge is prone to going out of date and is nontrivial to update.
1 code implementation • 21 May 2023 • Ge Gao, Hung-Ting Chen, Yoav Artzi, Eunsol Choi
We study continually improving an extractive question answering (QA) system via human user feedback.
1 code implementation • 19 May 2023 • Jifan Chen, Grace Kim, Aniruddh Sriram, Greg Durrett, Eunsol Choi
Evidence retrieval is a core part of automatic fact-checking.
1 code implementation • 2 May 2023 • Yasumasa Onoe, Michael J. Q. Zhang, Shankar Padmanabhan, Greg Durrett, Eunsol Choi
Pre-trained language models (LMs) are used for knowledge intensive tasks like question answering, but their knowledge gets continuously outdated as the world changes.
no code implementations • 1 Mar 2023 • Jeremy R. Cole, Palak Jain, Julian Martin Eisenschlos, Michael J. Q. Zhang, Eunsol Choi, Bhuwan Dhingra
We propose representing factual changes between paired documents as question-answer pairs, where the answer to the same question differs between two versions.
no code implementations • 22 Jan 2023 • Christopher Mohri, Daniel Andor, Eunsol Choi, Michael Collins
We study the problem of classification with a reject option for a fixed predictor, applicable in natural language processing.
no code implementations • 22 Dec 2022 • Xuewen Yao, Miriam Mikhelson, Megan Micheletti, Eunsol Choi, S Craig Watkins, Edison Thomaz, Kaya de Barbaro
In the current work, we provide a descriptive analysis of the concerns, psychological states, and motivations shared by healthy and distressed postpartum support seekers on two digital platforms, a one-on-one digital helpline and a publicly available online forum.
no code implementations • 2 Dec 2022 • Anuj Diwan, Ching-Feng Yeh, Wei-Ning Hsu, Paden Tomasello, Eunsol Choi, David Harwath, Abdelrahman Mohamed
Additionally, current speech recognition models and continual learning algorithms are not optimized to be compute-efficient.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+3
1 code implementation • 29 Nov 2022 • Anirudh Srinivasan, Eunsol Choi
We study politeness phenomena in nine typologically diverse languages.
no code implementations • 28 Nov 2022 • Xinyan Velocity Yu, Akari Asai, Trina Chatterjee, Junjie Hu, Eunsol Choi
While the NLP community is generally aware of resource disparities among languages, we lack research that quantifies the extent and types of such disparity.
1 code implementation • 1 Nov 2022 • Anuj Diwan, Layne Berry, Eunsol Choi, David Harwath, Kyle Mahowald
Recent visuolinguistic pre-trained models show promising progress on various end tasks such as image retrieval and video captioning.
no code implementations • 25 Oct 2022 • Hung-Ting Chen, Michael J. Q. Zhang, Eunsol Choi
Question answering models can use rich knowledge sources -- up to one hundred retrieved passages and parametric knowledge in the large-scale language model (LM).
no code implementations • NAACL (MIA) 2022 • Akari Asai, Shayne Longpre, Jungo Kasai, Chia-Hsuan Lee, Rui Zhang, Junjie Hu, Ikuya Yamada, Jonathan H. Clark, Eunsol Choi
We present the results of the Workshop on Multilingual Information Access (MIA) 2022 Shared Task, evaluating cross-lingual open-retrieval question answering (QA) systems in 16 typologically diverse languages.
no code implementations • 29 Jun 2022 • Venelin Kovatchev, Trina Chatterjee, Venkata S Govindarajan, Jifan Chen, Eunsol Choi, Gabriella Chronis, Anubrata Das, Katrin Erk, Matthew Lease, Junyi Jessy Li, Yating Wu, Kyle Mahowald
Developing methods to adversarially challenge NLP systems is a promising avenue for improving both model performance and interpretability.
no code implementations • NAACL 2022 • Shufan Wang, Fangyuan Xu, Laure Thompson, Eunsol Choi, Mohit Iyyer
We show that not only do state-of-the-art LFQA models struggle to generate relevant examples, but also that standard evaluation metrics such as ROUGE are insufficient to judge exemplification quality.
no code implementations • 14 May 2022 • Jifan Chen, Aniruddh Sriram, Eunsol Choi, Greg Durrett
Verifying complex political claims is a challenging task, especially when politicians use various tactics to subtly misrepresent the facts.
no code implementations • Findings (NAACL) 2022 • Yasumasa Onoe, Michael J. Q. Zhang, Eunsol Choi, Greg Durrett
Given its wide coverage on entity knowledge and temporal indexing, our dataset can be used to evaluate LMs and techniques designed to modify or extend their knowledge.
1 code implementation • ACL 2022 • Fangyuan Xu, Junyi Jessy Li, Eunsol Choi
Long-form answers, consisting of multiple sentences, can provide nuanced and comprehensive answers to a broader set of questions.
1 code implementation • ACL 2022 • Ge Gao, Eunsol Choi, Yoav Artzi
We study learning from user feedback for extractive question answering by simulating feedback using supervised data.
1 code implementation • EMNLP 2021 • Michael J. Q. Zhang, Eunsol Choi
To construct SituatedQA, we first identify such questions in existing QA datasets.
1 code implementation • EMNLP 2021 • Shujian Zhang, Chengyue Gong, Eunsol Choi
Introducing such multi label examples at the cost of annotating fewer examples brings clear gains on natural language inference task and entity typing task, even when we simply first train with a single label data and then fine tune with multi label examples.
2 code implementations • 3 Sep 2021 • Yasumasa Onoe, Michael J. Q. Zhang, Eunsol Choi, Greg Durrett
We introduce CREAK, a testbed for commonsense reasoning about entity knowledge, bridging fact-checking about entities (Harry Potter is a wizard and is skilled at riding a broomstick) with commonsense inferences (if you're good at a skill you can teach others how to do it).
1 code implementation • Findings (ACL) 2021 • Shujian Zhang, Chengyue Gong, Eunsol Choi
We study calibration in question answering, estimating whether model correctly predicts answer for each question.
1 code implementation • 18 Apr 2021 • Jifan Chen, Eunsol Choi, Greg Durrett
To build robust question answering systems, we need the ability to verify whether answers to questions are truly correct, not just "good enough" in the context of imperfect QA datasets.
2 code implementations • 18 Apr 2021 • Saadia Gabriel, Skyler Hallinan, Maarten Sap, Pemi Nguyen, Franziska Roesner, Eunsol Choi, Yejin Choi
We propose Misinfo Reaction Frames (MRF), a pragmatic formalism for modeling how readers might react to a news headline.
no code implementations • 13 Feb 2021 • Shujian Zhang, Chengyue Gong, Eunsol Choi
We depart from the standard practice of collecting a single reference per each training example, and find that collecting multiple references can achieve better accuracy under the fixed annotation budget.
no code implementations • 9 Feb 2021 • Eunsol Choi, Jennimaria Palomaki, Matthew Lamm, Tom Kwiatkowski, Dipanjan Das, Michael Collins
Models for question answering, dialogue agents, and summarization often interpret the meaning of a sentence in a rich context and use that meaning in a new context.
no code implementations • 1 Jan 2021 • Sewon Min, Jordan Boyd-Graber, Chris Alberti, Danqi Chen, Eunsol Choi, Michael Collins, Kelvin Guu, Hannaneh Hajishirzi, Kenton Lee, Jennimaria Palomaki, Colin Raffel, Adam Roberts, Tom Kwiatkowski, Patrick Lewis, Yuxiang Wu, Heinrich Küttler, Linqing Liu, Pasquale Minervini, Pontus Stenetorp, Sebastian Riedel, Sohee Yang, Minjoon Seo, Gautier Izacard, Fabio Petroni, Lucas Hosseini, Nicola De Cao, Edouard Grave, Ikuya Yamada, Sonse Shimaoka, Masatoshi Suzuki, Shumpei Miyawaki, Shun Sato, Ryo Takahashi, Jun Suzuki, Martin Fajcik, Martin Docekal, Karel Ondrej, Pavel Smrz, Hao Cheng, Yelong Shen, Xiaodong Liu, Pengcheng He, Weizhu Chen, Jianfeng Gao, Barlas Oguz, Xilun Chen, Vladimir Karpukhin, Stan Peshterliev, Dmytro Okhonko, Michael Schlichtkrull, Sonal Gupta, Yashar Mehdad, Wen-tau Yih
We review the EfficientQA competition from NeurIPS 2020.
no code implementations • ACL 2021 • Akari Asai, Eunsol Choi
However, datasets containing information-seeking queries where evidence documents are provided after the queries are written independently remain challenging.
3 code implementations • NAACL 2021 • Akari Asai, Jungo Kasai, Jonathan H. Clark, Kenton Lee, Eunsol Choi, Hannaneh Hajishirzi
Multilingual question answering tasks typically assume answers exist in the same language as the question.
1 code implementation • 8 Sep 2020 • Matthew Lamm, Jennimaria Palomaki, Chris Alberti, Daniel Andor, Eunsol Choi, Livio Baldini Soares, Michael Collins
A question answering system that in addition to providing an answer provides an explanation of the reasoning that leads to that answer has potential advantages in terms of debuggability, extensibility and trust.
1 code implementation • EMNLP 2020 • Thibault Févry, Livio Baldini Soares, Nicholas FitzGerald, Eunsol Choi, Tom Kwiatkowski
We introduce a new model - Entities as Experts (EAE) - that can access distinct memories of the entities mentioned in a piece of text.
2 code implementations • TACL 2020 • Jonathan H. Clark, Eunsol Choi, Michael Collins, Dan Garrette, Tom Kwiatkowski, Vitaly Nikolaev, Jennimaria Palomaki
Confidently making progress on multilingual modeling requires challenging, trustworthy evaluations.
1 code implementation • WS 2019 • Adam Fisch, Alon Talmor, Robin Jia, Minjoon Seo, Eunsol Choi, Danqi Chen
We present the results of the Machine Reading for Question Answering (MRQA) 2019 shared task on evaluating the generalization capabilities of reading comprehension systems.
1 code implementation • NAACL 2019 • Xiaochuang Han, Eunsol Choi, Chenhao Tan
Understanding the dynamics of international politics is important yet challenging for civilians.
3 code implementations • NAACL 2019 • Mandar Joshi, Eunsol Choi, Omer Levy, Daniel S. Weld, Luke Zettlemoyer
Reasoning about implied relationships (e. g., paraphrastic, common sense, encyclopedic) between pairs of words is crucial for many cross-sentence inference problems.
1 code implementation • ICLR 2019 • Hsin-Yuan Huang, Eunsol Choi, Wen-tau Yih
Conversational machine comprehension requires the understanding of the conversation history, such as previous question/answer pairs, the document context, and the current question.
Ranked #1 on
Question Answering
on QuAC
no code implementations • EMNLP 2018 • Eunsol Choi, He He, Mohit Iyyer, Mark Yatskar, Wen-tau Yih, Yejin Choi, Percy Liang, Luke Zettlemoyer
We present QuAC, a dataset for Question Answering in Context that contains 14K information-seeking QA dialogs (100K questions in total).
1 code implementation • EMNLP 2018 • Ge Gao, Eunsol Choi, Yejin Choi, Luke Zettlemoyer
We present end-to-end neural models for detecting metaphorical word use in context.
no code implementations • 21 Aug 2018 • Eunsol Choi, He He, Mohit Iyyer, Mark Yatskar, Wen-tau Yih, Yejin Choi, Percy Liang, Luke Zettlemoyer
We present QuAC, a dataset for Question Answering in Context that contains 14K information-seeking QA dialogs (100K questions in total).
1 code implementation • ACL 2018 • Eunsol Choi, Omer Levy, Yejin Choi, Luke Zettlemoyer
We introduce a new entity typing task: given a sentence with an entity mention, the goal is to predict a set of free-form phrases (e. g. skyscraper, songwriter, or criminal) that describe appropriate types for the target entity.
Ranked #4 on
Entity Typing
on Ontonotes v5 (English)
no code implementations • EMNLP 2017 • Hannah Rashkin, Eunsol Choi, Jin Yea Jang, Svitlana Volkova, Yejin Choi
We present an analytic study on the language of news media in the context of political fact-checking and fake news detection.
no code implementations • ACL 2017 • Eunsol Choi, Daniel Hewlett, Jakob Uszkoreit, Illia Polosukhin, Alex Lacoste, re, Jonathan Berant
We present a framework for question answering that can efficiently scale to longer documents while maintaining or even improving performance of state-of-the-art models.
2 code implementations • CONLL 2017 • Omer Levy, Minjoon Seo, Eunsol Choi, Luke Zettlemoyer
We show that relation extraction can be reduced to answering simple reading comprehension questions, by associating one or more natural-language questions with each relation slot.
3 code implementations • ACL 2017 • Mandar Joshi, Eunsol Choi, Daniel S. Weld, Luke Zettlemoyer
We present TriviaQA, a challenging reading comprehension dataset containing over 650K question-answer-evidence triples.
no code implementations • 6 Nov 2016 • Eunsol Choi, Daniel Hewlett, Alexandre Lacoste, Illia Polosukhin, Jakob Uszkoreit, Jonathan Berant
We present a framework for question answering that can efficiently scale to longer documents while maintaining or even improving performance of state-of-the-art models.
no code implementations • LREC 2016 • Eunsol Choi, Matic Horvat, Jonathan May, Kevin Knight, Daniel Marcu
Understanding the experimental results of a scientific paper is crucial to understanding its contribution and to comparing it with related work.