Search Results for author: Rachel Rudinger

Found 44 papers, 22 papers with code

On Measuring Social Biases in Sentence Encoders

1 code implementation NAACL 2019 Chandler May, Alex Wang, Shikha Bordia, Samuel R. Bowman, Rachel Rudinger

The Word Embedding Association Test shows that GloVe and word2vec word embeddings exhibit human-like implicit biases based on gender, race, and other social constructs (Caliskan et al., 2017).

Sentence Word Embeddings

Neural-Davidsonian Semantic Proto-role Labeling

1 code implementation EMNLP 2018 Rachel Rudinger, Adam Teichert, Ryan Culkin, Sheng Zhang, Benjamin Van Durme

We present a model for semantic proto-role labeling (SPRL) using an adapted bidirectional LSTM encoding strategy that we call "Neural-Davidsonian": predicate-argument structure is represented as pairs of hidden states corresponding to predicate and argument head tokens of the input sequence.

Attribute

Neural models of factuality

1 code implementation NAACL 2018 Rachel Rudinger, Aaron Steven White, Benjamin Van Durme

We present two neural models for event factuality prediction, which yield significant performance gains over previous models on three event factuality datasets: FactBank, UW, and MEANTIME.

Human Schema Curation via Causal Association Rule Mining

1 code implementation LREC (LAW) 2022 Noah Weber, Anton Belyy, Nils Holzenberger, Rachel Rudinger, Benjamin Van Durme

Event schemas are structured knowledge sources defining typical real-world scenarios (e. g., going to an airport).

"You are grounded!": Latent Name Artifacts in Pre-trained Language Models

1 code implementation6 Apr 2020 Vered Shwartz, Rachel Rudinger, Oyvind Tafjord

Pre-trained language models (LMs) may perpetuate biases originating in their training corpus to downstream models.

Reading Comprehension

SODAPOP: Open-Ended Discovery of Social Biases in Social Commonsense Reasoning Models

1 code implementation13 Oct 2022 Haozhe An, Zongxia Li, Jieyu Zhao, Rachel Rudinger

A common limitation of diagnostic tests for detecting social biases in NLP models is that they may only detect stereotypic associations that are pre-specified by the designer of the test.

Language Modelling Question Answering

Social Bias in Elicited Natural Language Inferences

1 code implementation WS 2017 Rachel Rudinger, Ch May, ler, Benjamin Van Durme

We analyze the Stanford Natural Language Inference (SNLI) corpus in an investigation of bias and stereotyping in NLP data.

Language Modelling Natural Language Inference +1

Theory-Grounded Measurement of U.S. Social Stereotypes in English Language Models

1 code implementation NAACL 2022 Yang Trista Cao, Anna Sotnikova, Hal Daumé III, Rachel Rudinger, Linda Zou

NLP models trained on text have been shown to reproduce human stereotypes, which can magnify harms to marginalized groups when systems are deployed at scale.

What do Large Language Models Learn about Scripts?

1 code implementation *SEM (NAACL) 2022 Abhilasha Sancheti, Rachel Rudinger

SIF is a two-staged framework that fine-tunes LM on a small set of ESD examples in the first stage.

It's Not Easy Being Wrong: Large Language Models Struggle with Process of Elimination Reasoning

1 code implementation13 Nov 2023 Nishant Balepur, Shramay Palta, Rachel Rudinger

Chain-of-thought (COT) prompting can help large language models (LLMs) reason toward correct answers, but its efficacy in reasoning toward incorrect answers is unexplored.

Multiple-choice

Artifacts or Abduction: How Do LLMs Answer Multiple-Choice Questions Without the Question?

1 code implementation19 Feb 2024 Nishant Balepur, Abhilasha Ravichander, Rachel Rudinger

We hope to motivate the use of stronger baselines in MCQA benchmarks, the design of robust MCQA datasets, and further efforts to explain LLM decision-making.

Decision Making Memorization +2

MedNLI Is Not Immune: Natural Language Inference Artifacts in the Clinical Domain

1 code implementation ACL 2021 Christine Herlihy, Rachel Rudinger

Crowdworker-constructed natural language inference (NLI) datasets have been found to contain statistical artifacts associated with the annotation process that allow hypothesis-only classifiers to achieve better-than-random performance (Poliak et al., 2018; Gururanganet et al., 2018; Tsuchiya, 2018).

Natural Language Inference Negation

Partial-input baselines show that NLI models can ignore context, but they don't

1 code implementation24 May 2022 Neha Srikanth, Rachel Rudinger

When strong partial-input baselines reveal artifacts in crowdsourced NLI datasets, the performance of full-input models trained on such datasets is often dismissed as reliance on spurious correlations.

Recognition of They/Them as Singular Personal Pronouns in Coreference Resolution

1 code implementation NAACL 2022 Connor Baumler, Rachel Rudinger

As using they/them as personal pronouns becomes increasingly common in English, it is important that coreference resolution systems work as well for individuals who use personal “they” as they do for those who use gendered personal pronouns.

coreference-resolution

Partial-input baselines show that NLI models can ignore context, but they don’t.

1 code implementation NAACL 2022 Neha Srikanth, Rachel Rudinger

When strong partial-input baselines reveal artifacts in crowdsourced NLI datasets, the performance of full-input models trained on such datasets is often dismissed as reliance on spurious correlations.

Multilingual large language models leak human stereotypes across language boundaries

1 code implementation12 Dec 2023 Yang Trista Cao, Anna Sotnikova, Jieyu Zhao, Linda X. Zou, Rachel Rudinger, Hal Daume III

We evaluate human stereotypes and stereotypical associations manifested in multilingual large language models such as mBERT, mT5, and ChatGPT.

Collecting Diverse Natural Language Inference Problems for Sentence Representation Evaluation

no code implementations EMNLP (ACL) 2018 Adam Poliak, Aparajita Haldar, Rachel Rudinger, J. Edward Hu, Ellie Pavlick, Aaron Steven White, Benjamin Van Durme

We present a large-scale collection of diverse natural language inference (NLI) datasets that help provide insight into how well a sentence representation captures distinct types of reasoning.

Natural Language Inference Sentence

Ordinal Common-sense Inference

no code implementations TACL 2017 Sheng Zhang, Rachel Rudinger, Kevin Duh, Benjamin Van Durme

Humans have the capacity to draw common-sense inferences from natural language: various things that are likely but not certain to hold based on established discourse, and are rarely stated explicitly.

Common Sense Reasoning Natural Language Inference

Computational linking theory

no code implementations8 Oct 2016 Aaron Steven White, Drew Reisinger, Rachel Rudinger, Kyle Rawlins, Benjamin Van Durme

A linking theory explains how verbs' semantic arguments are mapped to their syntactic arguments---the inverse of the Semantic Role Labeling task from the shallow semantic parsing literature.

Semantic Parsing Semantic Role Labeling

Lexicosyntactic Inference in Neural Models

no code implementations EMNLP 2018 Aaron Steven White, Rachel Rudinger, Kyle Rawlins, Benjamin Van Durme

We use this dataset, which we make publicly available, to probe the behavior of current state-of-the-art neural systems, showing that these systems make certain systematic errors that are clearly visible through the lens of factuality prediction.

Cross-lingual Decompositional Semantic Parsing

no code implementations EMNLP 2018 Sheng Zhang, Xutai Ma, Rachel Rudinger, Kevin Duh, Benjamin Van Durme

We introduce the task of cross-lingual decompositional semantic parsing: mapping content provided in a source language into a decompositional semantic analysis based on a target language.

Semantic Parsing

Semantic Proto-Roles

no code implementations TACL 2015 Drew Reisinger, Rachel Rudinger, Francis Ferraro, Craig Harman, Kyle Rawlins, Benjamin Van Durme

We present the first large-scale, corpus based verification of Dowty{'}s seminal theory of proto-roles.

Semantic Role Labeling

Causal Inference of Script Knowledge

no code implementations EMNLP 2020 Noah Weber, Rachel Rudinger, Benjamin Van Durme

When does a sequence of events define an everyday scenario and how can this knowledge be induced from text?

Causal Inference

``You are grounded!'': Latent Name Artifacts in Pre-trained Language Models

no code implementations EMNLP 2020 Vered Shwartz, Rachel Rudinger, Oyvind Tafjord

Pre-trained language models (LMs) may perpetuate biases originating in their training corpus to downstream models.

Reading Comprehension

Learning to Rationalize for Nonmonotonic Reasoning with Distant Supervision

no code implementations14 Dec 2020 Faeze Brahman, Vered Shwartz, Rachel Rudinger, Yejin Choi

In this paper, we investigate the extent to which neural models can reason about natural language rationales that explain model predictions, relying only on distant supervision with no additional annotation cost for human-written rationales.

Entailment Relation Aware Paraphrase Generation

no code implementations20 Mar 2022 Abhilasha Sancheti, Balaji Vasan Srinivasan, Rachel Rudinger

We introduce a new task of entailment relation aware paraphrase generation which aims at generating a paraphrase conforming to a given entailment relation (e. g. equivalent, forward entailing, or reverse entailing) with respect to a given input.

Natural Language Inference Paraphrase Generation +3

Agent-Specific Deontic Modality Detection in Legal Language

no code implementations23 Nov 2022 Abhilasha Sancheti, Aparna Garimella, Balaji Vasan Srinivasan, Rachel Rudinger

Legal documents are typically long and written in legalese, which makes it particularly difficult for laypeople to understand their rights and duties.

Natural Language Understanding Transfer Learning

What to Read in a Contract? Party-Specific Summarization of Legal Obligations, Entitlements, and Prohibitions

no code implementations19 Dec 2022 Abhilasha Sancheti, Aparna Garimella, Balaji Vasan Srinivasan, Rachel Rudinger

In this work, we propose a new task of party-specific extractive summarization for legal contracts to facilitate faster reviewing and improved comprehension of rights and duties.

Extractive Summarization Sentence +1

Nichelle and Nancy: The Influence of Demographic Attributes and Tokenization Length on First Name Biases

no code implementations26 May 2023 Haozhe An, Rachel Rudinger

We find that demographic attributes of a name (race, ethnicity, and gender) and name tokenization length are both factors that systematically affect the behavior of social commonsense reasoning models.

How often are errors in natural language reasoning due to paraphrastic variability?

no code implementations17 Apr 2024 Neha Srikanth, Marine Carpuat, Rachel Rudinger

We propose a metric for evaluating the paraphrastic consistency of natural language reasoning models based on the probability of a model achieving the same correctness on two paraphrases of the same problem.

Natural Language Inference

Cannot find the paper you are looking for? You can Submit a new open access paper.