Search Results for author: William Held

Found 13 papers, 7 papers with code

Social Skill Training with Large Language Models

no code implementations5 Apr 2024 Diyi Yang, Caleb Ziems, William Held, Omar Shaikh, Michael S. Bernstein, John Mitchell

People rely on social skills like conflict resolution to communicate effectively and to thrive in both work and personal life.

Unintended Impacts of LLM Alignment on Global Representation

no code implementations22 Feb 2024 Michael J. Ryan, William Held, Diyi Yang

Before being deployed for user-facing applications, developers align Large Language Models (LLMs) to user preferences through a variety of procedures, such as Reinforcement Learning From Human Feedback (RLHF) and Direct Preference Optimization (DPO).

Instruction Following

A Material Lens on Coloniality in NLP

no code implementations14 Nov 2023 William Held, Camille Harris, Michael Best, Diyi Yang

Coloniality, the continuation of colonial harms beyond "official" colonization, has pervasive effects across society and scientific fields.

Task-Agnostic Low-Rank Adapters for Unseen English Dialects

1 code implementation2 Nov 2023 Zedian Xiao, William Held, Yanchen Liu, Diyi Yang

Large Language Models (LLMs) are trained on corpora disproportionally weighted in favor of Standard American English.

Modeling Cross-Cultural Pragmatic Inference with Codenames Duet

1 code implementation4 Jun 2023 Omar Shaikh, Caleb Ziems, William Held, Aryan J. Pariani, Fred Morstatter, Diyi Yang

Prior work uses simple reference games to test models of pragmatic reasoning, often with unidentified speakers and listeners.

DADA: Dialect Adaptation via Dynamic Aggregation of Linguistic Rules

1 code implementation22 May 2023 Yanchen Liu, William Held, Diyi Yang

We show that DADA is effective for both single task and instruction finetuned language models, offering an extensible and interpretable framework for adapting existing LLMs to different English dialects.

Dialect Identification

Can Large Language Models Transform Computational Social Science?

1 code implementation12 Apr 2023 Caleb Ziems, William Held, Omar Shaikh, Jiaao Chen, Zhehao Zhang, Diyi Yang

We conclude that the performance of today's LLMs can augment the CSS research pipeline in two ways: (1) serving as zero-shot data annotators on human annotation teams, and (2) bootstrapping challenging creative generation tasks (e. g., explaining the underlying attributes of a text).

Persuasiveness

DAMP: Doubly Aligned Multilingual Parser for Task-Oriented Dialogue

1 code implementation15 Dec 2022 William Held, Christopher Hidey, Fei Liu, Eric Zhu, Rahul Goel, Diyi Yang, Rushin Shah

Modern virtual assistants use internal semantic parsing engines to convert user utterances to actionable commands.

Semantic Parsing XLM-R

On Second Thought, Let's Not Think Step by Step! Bias and Toxicity in Zero-Shot Reasoning

1 code implementation15 Dec 2022 Omar Shaikh, Hongxin Zhang, William Held, Michael Bernstein, Diyi Yang

Generating a Chain of Thought (CoT) has been shown to consistently improve large language model (LLM) performance on a wide range of NLP tasks.

Instruction Following Language Modelling +2

Shapley Head Pruning: Identifying and Removing Interference in Multilingual Transformers

no code implementations11 Oct 2022 William Held, Diyi Yang

However, as a fixed-size model acquires more languages, its performance across all languages degrades, a phenomenon termed interference.

Sentence Sentence Classification

Focus on what matters: Applying Discourse Coherence Theory to Cross Document Coreference

1 code implementation EMNLP 2021 William Held, Dan Iter, Dan Jurafsky

We model the entities/events in a reader's focus as a neighborhood within a learned latent embedding space which minimizes the distance between mentions and the centroids of their gold coreference clusters.

coreference-resolution Entity Cross-Document Coreference Resolution +2

The Effectiveness of Simple Hybrid Systems for Hypernym Discovery

no code implementations ACL 2019 William Held, Nizar Habash

Hypernymy modeling has largely been separated according to two paradigms, pattern-based methods and distributional methods.

Hypernym Discovery

Cannot find the paper you are looking for? You can Submit a new open access paper.