Search Results for author: Eunsol Choi

Found 80 papers, 40 papers with code

Misinfo Reaction Frames: Reasoning about Readers’ Reactions to News Headlines

2 code implementations ACL 2022 Saadia Gabriel, Skyler Hallinan, Maarten Sap, Pemi Nguyen, Franziska Roesner, Eunsol Choi, Yejin Choi

Even to a simple and short news headline, readers react in a multitude of ways: cognitively (e. g. inferring the writer’s intent), emotionally (e. g. feeling distrust), and behaviorally (e. g. sharing the news with their friends).

Misinformation

Can NLI Models Verify QA Systems’ Predictions?

1 code implementation Findings (EMNLP) 2021 Jifan Chen, Eunsol Choi, Greg Durrett

To build robust question answering systems, we need the ability to verify whether answers to questions are truly correct, not just “good enough” in the context of imperfect QA datasets.

Natural Language Inference Question Answering +1

Future of Information Retrieval Research in the Age of Generative AI

no code implementations3 Dec 2024 James Allan, Eunsol Choi, Daniel P. Lopresti, Hamed Zamani

In the fast-evolving field of information retrieval (IR), the integration of generative AI technologies such as large language models (LLMs) is transforming how users search for and interact with information.

Information Retrieval Retrieval

Recycled Attention: Efficient inference for long-context language models

no code implementations8 Nov 2024 Fangyuan Xu, Tanya Goyal, Eunsol Choi

Generating long sequences of tokens given a long-context input imposes a heavy computational burden for large language models (LLMs).

Language Modeling Language Modelling

RARe: Retrieval Augmented Retrieval with In-Context Examples

1 code implementation26 Oct 2024 Atula Tejaswi, Yoonsang Lee, Sujay Sanghavi, Eunsol Choi

Our approach, RARe, finetunes a pre-trained model with in-context examples whose query is semantically similar to the target query.

Decoder Domain Generalization +2

Diverging Preferences: When do Annotators Disagree and do Models Know?

no code implementations18 Oct 2024 Michael JQ Zhang, Zhilin Wang, Jena D. Hwang, Yi Dong, Olivier Delalleau, Yejin Choi, Eunsol Choi, Xiang Ren, Valentina Pyatkin

We find that the majority of disagreements are in opposition with standard reward modeling approaches, which are designed with the assumption that annotator disagreement is noise.

Modeling Future Conversation Turns to Teach LLMs to Ask Clarifying Questions

no code implementations17 Oct 2024 Michael J. Q. Zhang, W. Bradley Knox, Eunsol Choi

This allows LLMs to learn to ask clarifying questions when it can generate responses that are tailored to each user interpretation in future turns.

Contrastive Learning to Improve Retrieval for Real-world Fact Checking

no code implementations7 Oct 2024 Aniruddh Sriram, Fangyuan Xu, Eunsol Choi, Greg Durrett

By leveraging the AVeriTeC dataset, which annotates subquestions for claims with human written answers from evidence documents, we fine-tune Contriever with a contrastive objective based on multiple training signals, including distillation from GPT-4, evaluating subquestion answers, and gold labels in the dataset.

Contrastive Learning Fact Checking +2

Open-World Evaluation for Retrieving Diverse Perspectives

no code implementations26 Sep 2024 Hung-Ting Chen, Eunsol Choi

On this data, retrievers paired with a corpus are evaluated to surface a document set that contains diverse perspectives.

Diversity Language Modeling +2

Long-Form Answers to Visual Questions from Blind and Low Vision People

no code implementations12 Aug 2024 Mina Huh, Fangyuan Xu, Yi-Hao Peng, Chongyan Chen, Hansika Murugu, Danna Gurari, Eunsol Choi, Amy Pavel

Vision language models can now generate long-form answers to questions about images - long-form visual question answers (LFVQA).

Visual Question Answering (VQA)

CodeUpdateArena: Benchmarking Knowledge Editing on API Updates

no code implementations8 Jul 2024 Zeyu Leo Liu, Shrey Pandit, Xi Ye, Eunsol Choi, Greg Durrett

An instance in our benchmark consists of a synthetic API function update paired with a program synthesis example that uses the updated functionality; our goal is to update an LLM to be able to solve this program synthesis example without providing documentation of the update at inference time.

Benchmarking knowledge editing +1

CaLMQA: Exploring culturally specific long-form question answering across 23 languages

1 code implementation25 Jun 2024 Shane Arora, Marzena Karpinska, Hung-Ting Chen, Ipsita Bhattacharjee, Mohit Iyyer, Eunsol Choi

To bridge this gap, we introduce CaLMQA, a collection of 1. 5K complex culturally specific questions spanning 23 languages and 51 culturally agnostic questions translated from English into 22 other languages.

Long Form Question Answering

From Distributional to Overton Pluralism: Investigating Large Language Model Alignment

1 code implementation25 Jun 2024 Thom Lake, Eunsol Choi, Greg Durrett

Alignment suppresses irrelevant and unhelpful content while shifting the output distribution toward longer responses that cover information spanning several responses from the base LLM, essentially presenting diverse information in a single response.

Diversity Language Modeling +2

Exploring Design Choices for Building Language-Specific LLMs

1 code implementation20 Jun 2024 Atula Tejaswi, Nilesh Gupta, Eunsol Choi

In this paper, we study building language-specific LLMs by adapting monolingual and multilingual LLMs.

Model Selection

SVFT: Parameter-Efficient Fine-Tuning with Singular Vectors

1 code implementation30 May 2024 Vijay Lingam, Atula Tejaswi, Aditya Vavre, Aneesh Shetty, Gautham Krishna Gudur, Joydeep Ghosh, Alex Dimakis, Eunsol Choi, Aleksandar Bojchevski, Sujay Sanghavi

Extensive experiments on language and vision benchmarks show that SVFT recovers up to 96% of full fine-tuning performance while training only 0. 006 to 0. 25% of parameters, outperforming existing methods that only recover up to 85% performance using 0. 03 to 0. 8% of the trainable parameter budget.

parameter-efficient fine-tuning

AmbigDocs: Reasoning across Documents on Different Entities under the Same Name

no code implementations18 Apr 2024 Yoonsang Lee, Xi Ye, Eunsol Choi

and a set of documents discussing different people named Michael Jordan, can LMs distinguish entity mentions to generate a cohesive answer to the question?

KIWI: A Dataset of Knowledge-Intensive Writing Instructions for Answering Research Questions

no code implementations6 Mar 2024 Fangyuan Xu, Kyle Lo, Luca Soldaini, Bailey Kuehl, Eunsol Choi, David Wadden

To evaluate the capabilities of current LLMs on this task, we construct KIWI, a dataset of knowledge-intensive writing instructions in the scientific domain.

Instruction Following

BAT: Learning to Reason about Spatial Sounds with Large Language Models

no code implementations2 Feb 2024 Zhisheng Zheng, Puyuan Peng, Ziyang Ma, Xie Chen, Eunsol Choi, David Harwath

By integrating Spatial-AST with LLaMA-2 7B model, BAT transcends standard Sound Event Localization and Detection (SELD) tasks, enabling the model to reason about the relationships between the sounds in its environment.

Event Detection Language Modelling +5

Clarify When Necessary: Resolving Ambiguity Through Interaction with LMs

no code implementations16 Nov 2023 Michael J. Q. Zhang, Eunsol Choi

In this work, we study such behavior in LMs by proposing a task-agnostic framework for resolving ambiguity by asking users clarifying questions.

Machine Translation Natural Language Inference +1

Crafting In-context Examples according to LMs' Parametric Knowledge

1 code implementation16 Nov 2023 Yoonsang Lee, Pranav Atreya, Xi Ye, Eunsol Choi

We perform analysis on three multi-answer question answering datasets, which allows us to further study answer set ordering strategies based on the LM's knowledge of each answer.

Hallucination In-Context Learning +3

Understanding Retrieval Augmentation for Long-Form Question Answering

no code implementations18 Oct 2023 Hung-Ting Chen, Fangyuan Xu, Shane A. Arora, Eunsol Choi

Our study provides new insights on how retrieval augmentation impacts long, knowledge-rich text generation of LMs.

Long Form Question Answering Retrieval +1

RECOMP: Improving Retrieval-Augmented LMs with Compression and Selective Augmentation

2 code implementations6 Oct 2023 Fangyuan Xu, Weijia Shi, Eunsol Choi

Retrieving documents and prepending them in-context at inference time improves performance of language model (LMs) on a wide range of tasks.

Language Modeling Language Modelling +2

Development and Evaluation of Three Chatbots for Postpartum Mood and Anxiety Disorders

no code implementations14 Aug 2023 Xuewen Yao, Miriam Mikhelson, S. Craig Watkins, Eunsol Choi, Edison Thomaz, Kaya de Barbaro

In collaboration with Postpartum Support International (PSI), a non-profit organization dedicated to supporting caregivers with postpartum mood and anxiety disorders, we developed three chatbots to provide context-specific empathetic support to postpartum caregivers, leveraging both rule-based and generative models.

Chatbot

Propagating Knowledge Updates to LMs Through Distillation

1 code implementation NeurIPS 2023 Shankar Padmanabhan, Yasumasa Onoe, Michael J. Q. Zhang, Greg Durrett, Eunsol Choi

Then, we update the model parameters so that the distribution of the LM (the student) matches the distribution of the LM conditioned on the definition (the teacher) on the transfer set.

knowledge editing Language Modelling

When to Use Efficient Self Attention? Profiling Text, Speech and Image Transformer Variants

1 code implementation14 Jun 2023 Anuj Diwan, Eunsol Choi, David Harwath

We present the first unified study of the efficiency of self-attention-based Transformer variants spanning text, speech and vision.

Concise Answers to Complex Questions: Summarization of Long-form Answers

1 code implementation30 May 2023 Abhilash Potluri, Fangyuan Xu, Eunsol Choi

Long-form question answering systems provide rich information by presenting paragraph-level answers, often containing optional background or auxiliary information.

Extractive Summarization Long Form Question Answering

A Critical Evaluation of Evaluations for Long-form Question Answering

1 code implementation29 May 2023 Fangyuan Xu, Yixiao Song, Mohit Iyyer, Eunsol Choi

We present a careful analysis of experts' evaluation, which focuses on new aspects such as the comprehensiveness of the answer.

Long Form Question Answering Text Generation

Textless Speech-to-Speech Translation With Limited Parallel Data

2 code implementations24 May 2023 Anuj Diwan, Anirudh Srinivasan, David Harwath, Eunsol Choi

We first pretrain a model on large-scale monolingual speech data, finetune it with a small amount of parallel speech data (20-60 hours), and lastly train with an unsupervised backtranslation objective.

Automatic Speech Recognition Denoising +6

Mitigating Temporal Misalignment by Discarding Outdated Facts

1 code implementation24 May 2023 Michael J. Q. Zhang, Eunsol Choi

While large language models are able to retain vast amounts of world knowledge seen during pretraining, such knowledge is prone to going out of date and is nontrivial to update.

Question Answering Retrieval +1

Continually Improving Extractive QA via Human Feedback

1 code implementation21 May 2023 Ge Gao, Hung-Ting Chen, Yoav Artzi, Eunsol Choi

We study continually improving an extractive question answering (QA) system via human user feedback.

Domain Adaptation Extractive Question-Answering +1

Can LMs Learn New Entities from Descriptions? Challenges in Propagating Injected Knowledge

1 code implementation2 May 2023 Yasumasa Onoe, Michael J. Q. Zhang, Shankar Padmanabhan, Greg Durrett, Eunsol Choi

Pre-trained language models (LMs) are used for knowledge intensive tasks like question answering, but their knowledge gets continuously outdated as the world changes.

Question Answering

DIFFQG: Generating Questions to Summarize Factual Changes

no code implementations1 Mar 2023 Jeremy R. Cole, Palak Jain, Julian Martin Eisenschlos, Michael J. Q. Zhang, Eunsol Choi, Bhuwan Dhingra

We propose representing factual changes between paired documents as question-answer pairs, where the answer to the same question differs between two versions.

Change Detection Question Generation +1

Learning to Reject with a Fixed Predictor: Application to Decontextualization

no code implementations22 Jan 2023 Christopher Mohri, Daniel Andor, Eunsol Choi, Michael Collins

We study the problem of classification with a reject option for a fixed predictor, applicable in natural language processing.

Understanding Postpartum Parents' Experiences via Two Digital Platforms

no code implementations22 Dec 2022 Xuewen Yao, Miriam Mikhelson, Megan Micheletti, Eunsol Choi, S Craig Watkins, Edison Thomaz, Kaya de Barbaro

In the current work, we provide a descriptive analysis of the concerns, psychological states, and motivations shared by healthy and distressed postpartum support seekers on two digital platforms, a one-on-one digital helpline and a publicly available online forum.

Descriptive Vocal Bursts Valence Prediction

Beyond Counting Datasets: A Survey of Multilingual Dataset Construction and Necessary Resources

no code implementations28 Nov 2022 Xinyan Velocity Yu, Akari Asai, Trina Chatterjee, Junjie Hu, Eunsol Choi

While the NLP community is generally aware of resource disparities among languages, we lack research that quantifies the extent and types of such disparity.

Why is Winoground Hard? Investigating Failures in Visuolinguistic Compositionality

1 code implementation1 Nov 2022 Anuj Diwan, Layne Berry, Eunsol Choi, David Harwath, Kyle Mahowald

Recent visuolinguistic pre-trained models show promising progress on various end tasks such as image retrieval and video captioning.

Data Augmentation Image Retrieval +2

Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence

no code implementations25 Oct 2022 Hung-Ting Chen, Michael J. Q. Zhang, Eunsol Choi

Question answering models can use rich knowledge sources -- up to one hundred retrieved passages and parametric knowledge in the large-scale language model (LM).

Language Modeling Language Modelling +2

MIA 2022 Shared Task: Evaluating Cross-lingual Open-Retrieval Question Answering for 16 Diverse Languages

no code implementations NAACL (MIA) 2022 Akari Asai, Shayne Longpre, Jungo Kasai, Chia-Hsuan Lee, Rui Zhang, Junjie Hu, Ikuya Yamada, Jonathan H. Clark, Eunsol Choi

We present the results of the Workshop on Multilingual Information Access (MIA) 2022 Shared Task, evaluating cross-lingual open-retrieval question answering (QA) systems in 16 typologically diverse languages.

Question Answering Retrieval

Modeling Exemplification in Long-form Question Answering via Retrieval

no code implementations NAACL 2022 Shufan Wang, Fangyuan Xu, Laure Thompson, Eunsol Choi, Mohit Iyyer

We show that not only do state-of-the-art LFQA models struggle to generate relevant examples, but also that standard evaluation metrics such as ROUGE are insufficient to judge exemplification quality.

Long Form Question Answering Retrieval

Generating Literal and Implied Subquestions to Fact-check Complex Claims

no code implementations14 May 2022 Jifan Chen, Aniruddh Sriram, Eunsol Choi, Greg Durrett

Verifying complex political claims is a challenging task, especially when politicians use various tactics to subtly misrepresent the facts.

Fact Checking

Entity Cloze By Date: What LMs Know About Unseen Entities

no code implementations Findings (NAACL) 2022 Yasumasa Onoe, Michael J. Q. Zhang, Eunsol Choi, Greg Durrett

Given its wide coverage on entity knowledge and temporal indexing, our dataset can be used to evaluate LMs and techniques designed to modify or extend their knowledge.

How Do We Answer Complex Questions: Discourse Structure of Long-form Answers

1 code implementation ACL 2022 Fangyuan Xu, Junyi Jessy Li, Eunsol Choi

Long-form answers, consisting of multiple sentences, can provide nuanced and comprehensive answers to a broader set of questions.

Natural Questions Sentence

Learning with Different Amounts of Annotation: From Zero to Many Labels

1 code implementation EMNLP 2021 Shujian Zhang, Chengyue Gong, Eunsol Choi

Introducing such multi label examples at the cost of annotating fewer examples brings clear gains on natural language inference task and entity typing task, even when we simply first train with a single label data and then fine tune with multi label examples.

Data Augmentation Entity Typing +1

CREAK: A Dataset for Commonsense Reasoning over Entity Knowledge

2 code implementations3 Sep 2021 Yasumasa Onoe, Michael J. Q. Zhang, Eunsol Choi, Greg Durrett

We introduce CREAK, a testbed for commonsense reasoning about entity knowledge, bridging fact-checking about entities (Harry Potter is a wizard and is skilled at riding a broomstick) with commonsense inferences (if you're good at a skill you can teach others how to do it).

Fact Checking Fact Verification +1

Can NLI Models Verify QA Systems' Predictions?

1 code implementation18 Apr 2021 Jifan Chen, Eunsol Choi, Greg Durrett

To build robust question answering systems, we need the ability to verify whether answers to questions are truly correct, not just "good enough" in the context of imperfect QA datasets.

Natural Language Inference Question Answering

Capturing Label Distribution: A Case Study in NLI

no code implementations13 Feb 2021 Shujian Zhang, Chengyue Gong, Eunsol Choi

We depart from the standard practice of collecting a single reference per each training example, and find that collecting multiple references can achieve better accuracy under the fixed annotation budget.

Natural Language Inference

Decontextualization: Making Sentences Stand-Alone

no code implementations9 Feb 2021 Eunsol Choi, Jennimaria Palomaki, Matthew Lamm, Tom Kwiatkowski, Dipanjan Das, Michael Collins

Models for question answering, dialogue agents, and summarization often interpret the meaning of a sentence in a rich context and use that meaning in a new context.

document understanding Question Answering +1

Challenges in Information-Seeking QA: Unanswerable Questions and Paragraph Retrieval

no code implementations ACL 2021 Akari Asai, Eunsol Choi

However, datasets containing information-seeking queries where evidence documents are provided after the queries are written independently remain challenging.

answerability prediction Language Modelling +4

QED: A Framework and Dataset for Explanations in Question Answering

1 code implementation8 Sep 2020 Matthew Lamm, Jennimaria Palomaki, Chris Alberti, Daniel Andor, Eunsol Choi, Livio Baldini Soares, Michael Collins

A question answering system that in addition to providing an answer provides an explanation of the reasoning that leads to that answer has potential advantages in terms of debuggability, extensibility and trust.

Explanation Generation Natural Questions +1

MRQA 2019 Shared Task: Evaluating Generalization in Reading Comprehension

1 code implementation WS 2019 Adam Fisch, Alon Talmor, Robin Jia, Minjoon Seo, Eunsol Choi, Danqi Chen

We present the results of the Machine Reading for Question Answering (MRQA) 2019 shared task on evaluating the generalization capabilities of reading comprehension systems.

Multi-Task Learning Question Answering +1

No Permanent Friends or Enemies: Tracking Relationships between Nations from News

1 code implementation NAACL 2019 Xiaochuang Han, Eunsol Choi, Chenhao Tan

Understanding the dynamics of international politics is important yet challenging for civilians.

pair2vec: Compositional Word-Pair Embeddings for Cross-Sentence Inference

3 code implementations NAACL 2019 Mandar Joshi, Eunsol Choi, Omer Levy, Daniel S. Weld, Luke Zettlemoyer

Reasoning about implied relationships (e. g., paraphrastic, common sense, encyclopedic) between pairs of words is crucial for many cross-sentence inference problems.

Common Sense Reasoning Sentence +1

FlowQA: Grasping Flow in History for Conversational Machine Comprehension

1 code implementation ICLR 2019 Hsin-Yuan Huang, Eunsol Choi, Wen-tau Yih

Conversational machine comprehension requires the understanding of the conversation history, such as previous question/answer pairs, the document context, and the current question.

Question Answering Reading Comprehension +1

QuAC: Question Answering in Context

no code implementations EMNLP 2018 Eunsol Choi, He He, Mohit Iyyer, Mark Yatskar, Wen-tau Yih, Yejin Choi, Percy Liang, Luke Zettlemoyer

We present QuAC, a dataset for Question Answering in Context that contains 14K information-seeking QA dialogs (100K questions in total).

Question Answering Reading Comprehension

Neural Metaphor Detection in Context

1 code implementation EMNLP 2018 Ge Gao, Eunsol Choi, Yejin Choi, Luke Zettlemoyer

We present end-to-end neural models for detecting metaphorical word use in context.

QuAC : Question Answering in Context

no code implementations21 Aug 2018 Eunsol Choi, He He, Mohit Iyyer, Mark Yatskar, Wen-tau Yih, Yejin Choi, Percy Liang, Luke Zettlemoyer

We present QuAC, a dataset for Question Answering in Context that contains 14K information-seeking QA dialogs (100K questions in total).

Question Answering Reading Comprehension

Ultra-Fine Entity Typing

1 code implementation ACL 2018 Eunsol Choi, Omer Levy, Yejin Choi, Luke Zettlemoyer

We introduce a new entity typing task: given a sentence with an entity mention, the goal is to predict a set of free-form phrases (e. g. skyscraper, songwriter, or criminal) that describe appropriate types for the target entity.

Entity Linking Entity Typing +1

Coarse-to-Fine Question Answering for Long Documents

no code implementations ACL 2017 Eunsol Choi, Daniel Hewlett, Jakob Uszkoreit, Illia Polosukhin, Alex Lacoste, re, Jonathan Berant

We present a framework for question answering that can efficiently scale to longer documents while maintaining or even improving performance of state-of-the-art models.

Question Answering Reading Comprehension +2

Zero-Shot Relation Extraction via Reading Comprehension

2 code implementations CONLL 2017 Omer Levy, Minjoon Seo, Eunsol Choi, Luke Zettlemoyer

We show that relation extraction can be reduced to answering simple reading comprehension questions, by associating one or more natural-language questions with each relation slot.

Reading Comprehension Relation +5

Hierarchical Question Answering for Long Documents

no code implementations6 Nov 2016 Eunsol Choi, Daniel Hewlett, Alexandre Lacoste, Illia Polosukhin, Jakob Uszkoreit, Jonathan Berant

We present a framework for question answering that can efficiently scale to longer documents while maintaining or even improving performance of state-of-the-art models.

Question Answering Reading Comprehension +2

Cannot find the paper you are looking for? You can Submit a new open access paper.