Search Results for author: Dan Jurafsky

Found 132 papers, 65 papers with code

Belief in the Machine: Investigating Epistemological Blind Spots of Language Models

1 code implementation28 Oct 2024 Mirac Suzgun, Tayfun Gur, Federico Bianchi, Daniel E. Ho, Thomas Icard, Dan Jurafsky, James Zou

These findings highlight significant concerns about current LMs' ability to reason about truth, belief, and knowledge while emphasizing the need for advancements in these areas before broad deployment in critical sectors.

Epistemic Reasoning Fact Checking +1

Bayesian scaling laws for in-context learning

1 code implementation21 Oct 2024 Aryaman Arora, Dan Jurafsky, Christopher Potts, Noah D. Goodman

In all cases, Bayesian scaling laws accurately predict the conditions under which ICL will cause the suppressed behavior to reemerge, which sheds light on the ineffectiveness of post-training at increasing LLM safety.

In-Context Learning Safety Alignment

Can Unconfident LLM Annotations Be Used for Confident Conclusions?

1 code implementation27 Aug 2024 Kristina Gligorić, Tijana Zrnic, Cinoo Lee, Emmanuel J. Candès, Dan Jurafsky

We introduce Confidence-Driven Inference: a method that combines LLM annotations and LLM confidence indicators to strategically select which human annotations should be collected, with the goal of producing accurate statistical estimates and provably valid confidence intervals while reducing the number of human annotations needed.

valid

A layer-wise analysis of Mandarin and English suprasegmentals in SSL speech models

no code implementations24 Aug 2024 Antón de la Fuente, Dan Jurafsky

This study asks how self-supervised speech models represent suprasegmental categories like Mandarin lexical tone, English lexical stress, and English phrasal accents.

Specificity

h4rm3l: A Dynamic Benchmark of Composable Jailbreak Attacks for LLM Safety Assessment

no code implementations9 Aug 2024 Moussa Koulako Bala Doumbouya, Ananjan Nandi, Gabriel Poesia, Davide Ghilardi, Anna Goldie, Federico Bianchi, Dan Jurafsky, Christopher D. Manning

The safety of Large Language Models (LLMs) remains a critical concern due to a lack of adequate benchmarks for systematically evaluating their ability to resist generating harmful content.

Benchmarking Program Synthesis

Data Checklist: On Unit-Testing Datasets with Usable Information

1 code implementation6 Aug 2024 Heidi C. Zhang, Shabnam Behzad, Kawin Ethayarajh, Dan Jurafsky

Model checklists (Ribeiro et al., 2020) have emerged as a useful tool for understanding the behavior of LLMs, analogous to unit-testing in software engineering.

Rel-A.I.: An Interaction-Centered Approach To Measuring Human-LM Reliance

no code implementations10 Jul 2024 Kaitlyn Zhou, Jena D. Hwang, Xiang Ren, Nouha Dziri, Dan Jurafsky, Maarten Sap

The ability to communicate uncertainty, risk, and limitation is crucial for the safety of large language models.

Sentence

ML-SUPERB 2.0: Benchmarking Multilingual Speech Models Across Modeling Constraints, Languages, and Datasets

no code implementations12 Jun 2024 Jiatong Shi, Shih-Heng Wang, William Chen, Martijn Bartelds, Vanya Bannihatti Kumar, Jinchuan Tian, Xuankai Chang, Dan Jurafsky, Karen Livescu, Hung-Yi Lee, Shinji Watanabe

This paper presents ML-SUPERB~2. 0, which is a new benchmark for evaluating pre-trained SSL and supervised speech models across downstream models, fine-tuning setups, and efficient model adaptation approaches.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +4

ReFT: Representation Finetuning for Language Models

2 code implementations4 Apr 2024 Zhengxuan Wu, Aryaman Arora, Zheng Wang, Atticus Geiger, Dan Jurafsky, Christopher D. Manning, Christopher Potts

We define a strong instance of the ReFT family, Low-rank Linear Subspace ReFT (LoReFT), and we identify an ablation of this method that trades some performance for increased efficiency.

Arithmetic Reasoning

NLP Systems That Can't Tell Use from Mention Censor Counterspeech, but Teaching the Distinction Helps

1 code implementation2 Apr 2024 Kristina Gligoric, Myra Cheng, Lucia Zheng, Esin Durmus, Dan Jurafsky

The use of words to convey speaker's intent is traditionally distinguished from the `mention' of words for quoting what someone said, or pointing out properties of a word.

Hate Speech Detection Misinformation

Dialect prejudice predicts AI decisions about people's character, employability, and criminality

1 code implementation1 Mar 2024 Valentin Hofmann, Pratyusha Ria Kalluri, Dan Jurafsky, Sharese King

Here, we demonstrate that language models embody covert racism in the form of dialect prejudice: we extend research showing that Americans hold raciolinguistic stereotypes about speakers of African American English and find that language models have the same prejudice, exhibiting covert stereotypes that are more negative than any human stereotypes about African Americans ever experimentally recorded, although closest to the ones from before the civil rights movement.

CausalGym: Benchmarking causal interpretability methods on linguistic tasks

1 code implementation19 Feb 2024 Aryaman Arora, Dan Jurafsky, Christopher Potts

Language models (LMs) have proven to be powerful tools for psycholinguistic research, but most prior work has focused on purely behavioural measures (e. g., surprisal comparisons).

Benchmarking Interpretability Techniques for Deep Learning

How Well Can LLMs Negotiate? NegotiationArena Platform and Analysis

1 code implementation8 Feb 2024 Federico Bianchi, Patrick John Chia, Mert Yuksekgonul, Jacopo Tagliabue, Dan Jurafsky, James Zou

We develop NegotiationArena: a flexible framework for evaluating and probing the negotiation abilities of LLM agents.

AnthroScore: A Computational Linguistic Measure of Anthropomorphism

1 code implementation3 Feb 2024 Myra Cheng, Kristina Gligoric, Tiziano Piccardi, Dan Jurafsky

Anthropomorphism, or the attribution of human-like characteristics to non-human entities, has shaped conversations about the impacts and possibilities of technology.

Language Modelling Misinformation

KTO: Model Alignment as Prospect Theoretic Optimization

3 code implementations2 Feb 2024 Kawin Ethayarajh, Winnie Xu, Niklas Muennighoff, Dan Jurafsky, Douwe Kiela

Kahneman & Tversky's $\textit{prospect theory}$ tells us that humans perceive random variables in a biased but well-defined manner (1992); for example, humans are famously loss-averse.

Attribute

Multilingual self-supervised speech representations improve the speech recognition of low-resource African languages with codeswitching

no code implementations25 Nov 2023 Tolúlopé Ògúnrèmí, Christopher D. Manning, Dan Jurafsky

While many speakers of low-resource languages regularly code-switch between their languages and other regional languages or English, datasets of codeswitched speech are too small to train bespoke acoustic models from scratch or do language model rescoring.

Language Modelling speech-recognition +1

Grounding Gaps in Language Model Generations

no code implementations15 Nov 2023 Omar Shaikh, Kristina Gligorić, Ashna Khetan, Matthias Gerstgrasser, Diyi Yang, Dan Jurafsky

To understand the roots of the identified grounding gap, we examine the role of instruction tuning and preference optimization, finding that training on contemporary preference data leads to a reduction in generated grounding acts.

Language Modelling

A Benchmark for Learning to Translate a New Language from One Grammar Book

no code implementations28 Sep 2023 Garrett Tanzer, Mirac Suzgun, Eline Visser, Dan Jurafsky, Luke Melas-Kyriazi

In this paper, we introduce MTOB (Machine Translation from One Book), a benchmark for learning to translate between English and Kalamang -- a language with less than 200 speakers and therefore virtually no presence on the web -- using several hundred pages of field linguistics reference materials.

In-Context Learning Machine Translation +1

Safety-Tuned LLaMAs: Lessons From Improving the Safety of Large Language Models that Follow Instructions

4 code implementations14 Sep 2023 Federico Bianchi, Mirac Suzgun, Giuseppe Attanasio, Paul Röttger, Dan Jurafsky, Tatsunori Hashimoto, James Zou

Training large language models to follow instructions makes them perform better on a wide range of tasks and generally become more helpful.

Learning the meanings of function words from grounded language using a visual question answering model

1 code implementation16 Aug 2023 Eva Portelance, Michael C. Frank, Dan Jurafsky

Furthermore, we find that these models can learn the meanings of logical connectives and and or without any prior knowledge of logical reasoning, as well as early evidence that they are sensitive to alternative expressions when interpreting language.

Logical Reasoning Question Answering +2

Othering and low status framing of immigrant cuisines in US restaurant reviews and large language models

1 code implementation14 Jul 2023 Yiwei Luo, Kristina Gligorić, Dan Jurafsky

Through careful linguistic analyses, we evaluate social theories about attitudes toward immigrant cuisine in a large-scale study of framing differences in 2. 1M English language Yelp reviews.

Text Generation

Ecosystem-level Analysis of Deployed Machine Learning Reveals Homogeneous Outcomes

no code implementations NeurIPS 2023 Connor Toups, Rishi Bommasani, Kathleen A. Creel, Sarah H. Bana, Dan Jurafsky, Percy Liang

In practice, the societal impact of machine learning is determined by the surrounding context of machine learning deployments.

Developing Speech Processing Pipelines for Police Accountability

no code implementations9 Jun 2023 Anjalie Field, Prateek Verma, Nay San, Jennifer L. Eberhardt, Dan Jurafsky

We investigate the potential of large pre-trained speech models for facilitating reviews, focusing on ASR and officer speech detection in footage from traffic stops.

Marked Personas: Using Natural Language Prompts to Measure Stereotypes in Language Models

1 code implementation29 May 2023 Myra Cheng, Esin Durmus, Dan Jurafsky

To recognize and mitigate harms from large language models (LLMs), we need to understand the prevalence and nuances of stereotypes in LLM outputs.

Story Generation

Making More of Little Data: Improving Low-Resource Automatic Speech Recognition Using Data Augmentation

1 code implementation18 May 2023 Martijn Bartelds, Nay San, Bradley McDonnell, Dan Jurafsky, Martijn Wieling

For Gronings, for which there was a pre-existing text-to-speech (TTS) system available, we also examined the use of TTS to generate ASR training data from text-only sources.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +3

string2string: A Modern Python Library for String-to-String Algorithms

2 code implementations27 Apr 2023 Mirac Suzgun, Stuart M. Shieber, Dan Jurafsky

It includes traditional algorithmic solutions as well as recent advanced neural approaches to tackle various problems in string alignment, distance measurement, lexical and semantic search, and similarity analysis -- along with several helpful visualization tools and metrics to facilitate the interpretation and analysis of these methods.

Injecting structural hints: Using language models to study inductive biases in language learning

1 code implementation25 Apr 2023 Isabel Papadimitriou, Dan Jurafsky

Our study leverages the capabilities of transformer models to run controlled language learning experiments that are not possible to run on humans, and surfaces hypotheses about the structures that facilitate language learning in both humans and machines.

Inductive Bias Transfer Learning

Navigating the Grey Area: How Expressions of Uncertainty and Overconfidence Affect Language Models

no code implementations26 Feb 2023 Kaitlyn Zhou, Dan Jurafsky, Tatsunori Hashimoto

The increased deployment of LMs for real-world tasks involving knowledge and facts makes it important to understand model epistemology: what LMs think they know, and how their attitudes toward that knowledge are affected by language use in their inputs.

Decision Making Question Answering +1

Leveraging supplementary text data to kick-start automatic speech recognition system development with limited transcriptions

no code implementations9 Feb 2023 Nay San, Martijn Bartelds, Blaine Billings, Ella de Falco, Hendi Feriza, Johan Safri, Wawan Sahrozi, Ben Foley, Bradley McDonnell, Dan Jurafsky

We perform experiments using 10 minutes of transcribed speech from English (for replicating prior work) and two additional pairs of languages differing in the availability of supplemental text data: Gronings and Frisian (~7. 5M token corpora available), and Besemah and Nasal (only small lexica available).

Automatic Speech Recognition Automatic Speech Recognition (ASR) +1

Self-Destructing Models: Increasing the Costs of Harmful Dual Uses of Foundation Models

1 code implementation27 Nov 2022 Peter Henderson, Eric Mitchell, Christopher D. Manning, Dan Jurafsky, Chelsea Finn

A growing ecosystem of large, open-source foundation models has reduced the labeled data and technical expertise necessary to apply machine learning to many new problems.

Blocking Meta-Learning

Picking on the Same Person: Does Algorithmic Monoculture lead to Outcome Homogenization?

no code implementations25 Nov 2022 Rishi Bommasani, Kathleen A. Creel, Ananya Kumar, Dan Jurafsky, Percy Liang

As the scope of machine learning broadens, we observe a recurring theme of algorithmic monoculture: the same systems, or systems that share components (e. g. training data), are deployed by multiple decision-makers.

Fairness

Follow the Wisdom of the Crowd: Effective Text Generation via Minimum Bayes Risk Decoding

1 code implementation14 Nov 2022 Mirac Suzgun, Luke Melas-Kyriazi, Dan Jurafsky

In open-ended natural-language generation, existing text decoding methods typically struggle to produce text which is both diverse and high-quality.

Diversity Style Transfer +1

Easily Accessible Text-to-Image Generation Amplifies Demographic Stereotypes at Large Scale

1 code implementation7 Nov 2022 Federico Bianchi, Pratyusha Kalluri, Esin Durmus, Faisal Ladhak, Myra Cheng, Debora Nozza, Tatsunori Hashimoto, Dan Jurafsky, James Zou, Aylin Caliskan

For example, we find cases of prompting for basic traits or social roles resulting in images reinforcing whiteness as ideal, prompting for occupations resulting in amplification of racial and gender disparities, and prompting for objects resulting in reification of American norms.

Text-to-Image Generation

Multilingual BERT has an accent: Evaluating English influences on fluency in multilingual models

no code implementations11 Oct 2022 Isabel Papadimitriou, Kezia Lopez, Dan Jurafsky

Here we show another problem with multilingual models: grammatical structures in higher-resource languages bleed into lower-resource languages, a phenomenon we call grammatical structure bias.

Language Modelling

When and why vision-language models behave like bags-of-words, and what to do about it?

1 code implementation4 Oct 2022 Mert Yuksekgonul, Federico Bianchi, Pratyusha Kalluri, Dan Jurafsky, James Zou

ARO consists of Visual Genome Attribution, to test the understanding of objects' properties; Visual Genome Relation, to test for relational understanding; and COCO & Flickr30k-Order, to test for order sensitivity.

Contrastive Learning Retrieval +1

Pile of Law: Learning Responsible Data Filtering from the Law and a 256GB Open-Source Legal Dataset

1 code implementation1 Jul 2022 Peter Henderson, Mark S. Krass, Lucia Zheng, Neel Guha, Christopher D. Manning, Dan Jurafsky, Daniel E. Ho

One concern with the rise of large language models lies with their potential for significant harm, particularly from pretraining on biased, obscene, copyrighted, and private information.

Prompt-and-Rerank: A Method for Zero-Shot and Few-Shot Arbitrary Textual Style Transfer with Small Language Models

1 code implementation23 May 2022 Mirac Suzgun, Luke Melas-Kyriazi, Dan Jurafsky

We propose a method for arbitrary textual style transfer (TST)--the task of transforming a text into any given style--utilizing general-purpose pre-trained language models.

Style Transfer

Richer Countries and Richer Representations

1 code implementation Findings (ACL) 2022 Kaitlyn Zhou, Kawin Ethayarajh, Dan Jurafsky

We examine whether some countries are more richly represented in embedding space than others.

Problems with Cosine as a Measure of Embedding Similarity for High Frequency Words

2 code implementations ACL 2022 Kaitlyn Zhou, Kawin Ethayarajh, Dallas Card, Dan Jurafsky

Cosine similarity of contextual embeddings is used in many NLP tasks (e. g., QA, IR, MT) and metrics (e. g., BERTScore).

Modular Domain Adaptation

1 code implementation Findings (ACL) 2022 Junshen K. Chen, Dallas Card, Dan Jurafsky

Off-the-shelf models are widely used by computational social science researchers to measure properties of text, such as sentiment.

Domain Adaptation text-classification +1

Focus on what matters: Applying Discourse Coherence Theory to Cross Document Coreference

1 code implementation EMNLP 2021 William Held, Dan Iter, Dan Jurafsky

We model the entities/events in a reader's focus as a neighborhood within a learned latent embedding space which minimizes the distance between mentions and the centroids of their gold coreference clusters.

coreference-resolution Entity Cross-Document Coreference Resolution +2

The Emergence of the Shape Bias Results from Communicative Efficiency

1 code implementation CoNLL (EMNLP) 2021 Eva Portelance, Michael C. Frank, Dan Jurafsky, Alessandro Sordoni, Romain Laroche

By the age of two, children tend to assume that new word categories are based on objects' shape, rather than their color or texture; this assumption is called the shape bias.

On the Opportunities and Risks of Foundation Models

2 code implementations16 Aug 2021 Rishi Bommasani, Drew A. Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael S. Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, Erik Brynjolfsson, Shyamal Buch, Dallas Card, Rodrigo Castellon, Niladri Chatterji, Annie Chen, Kathleen Creel, Jared Quincy Davis, Dora Demszky, Chris Donahue, Moussa Doumbouya, Esin Durmus, Stefano Ermon, John Etchemendy, Kawin Ethayarajh, Li Fei-Fei, Chelsea Finn, Trevor Gale, Lauren Gillespie, Karan Goel, Noah Goodman, Shelby Grossman, Neel Guha, Tatsunori Hashimoto, Peter Henderson, John Hewitt, Daniel E. Ho, Jenny Hong, Kyle Hsu, Jing Huang, Thomas Icard, Saahil Jain, Dan Jurafsky, Pratyusha Kalluri, Siddharth Karamcheti, Geoff Keeling, Fereshte Khani, Omar Khattab, Pang Wei Koh, Mark Krass, Ranjay Krishna, Rohith Kuditipudi, Ananya Kumar, Faisal Ladhak, Mina Lee, Tony Lee, Jure Leskovec, Isabelle Levent, Xiang Lisa Li, Xuechen Li, Tengyu Ma, Ali Malik, Christopher D. Manning, Suvir Mirchandani, Eric Mitchell, Zanele Munyikwa, Suraj Nair, Avanika Narayan, Deepak Narayanan, Ben Newman, Allen Nie, Juan Carlos Niebles, Hamed Nilforoshan, Julian Nyarko, Giray Ogut, Laurel Orr, Isabel Papadimitriou, Joon Sung Park, Chris Piech, Eva Portelance, Christopher Potts, aditi raghunathan, Rob Reich, Hongyu Ren, Frieda Rong, Yusuf Roohani, Camilo Ruiz, Jack Ryan, Christopher Ré, Dorsa Sadigh, Shiori Sagawa, Keshav Santhanam, Andy Shih, Krishnan Srinivasan, Alex Tamkin, Rohan Taori, Armin W. Thomas, Florian Tramèr, Rose E. Wang, William Wang, Bohan Wu, Jiajun Wu, Yuhuai Wu, Sang Michael Xie, Michihiro Yasunaga, Jiaxuan You, Matei Zaharia, Michael Zhang, Tianyi Zhang, Xikun Zhang, Yuhui Zhang, Lucia Zheng, Kaitlyn Zhou, Percy Liang

AI is undergoing a paradigm shift with the rise of models (e. g., BERT, DALL-E, GPT-3) that are trained on broad data at scale and are adaptable to a wide range of downstream tasks.

Transfer Learning

Measuring Conversational Uptake: A Case Study on Student-Teacher Interactions

1 code implementation ACL 2021 Dorottya Demszky, Jing Liu, Zid Mancenido, Julie Cohen, Heather Hill, Dan Jurafsky, Tatsunori Hashimoto

In conversation, uptake happens when a speaker builds on the contribution of their interlocutor by, for example, acknowledging, repeating or reformulating what they have said.

Math Question Answering

Attention Flows are Shapley Value Explanations

no code implementations ACL 2021 Kawin Ethayarajh, Dan Jurafsky

Shapley Values, a solution to the credit assignment problem in cooperative game theory, are a popular type of explanation in machine learning, having been used to explain the importance of features, embeddings, and even neurons.

Sensitivity as a Complexity Measure for Sequence Classification Tasks

1 code implementation21 Apr 2021 Michael Hahn, Dan Jurafsky, Richard Futrell

We introduce a theoretical framework for understanding and predicting the complexity of sequence classification tasks, using a novel extension of the theory of Boolean function sensitivity.

General Classification text-classification +1

Frequency-based Distortions in Contextualized Word Embeddings

no code implementations17 Apr 2021 Kaitlyn Zhou, Kawin Ethayarajh, Dan Jurafsky

How does word frequency in pre-training data affect the behavior of similarity metrics in contextualized BERT embeddings?

Semantic Similarity Semantic Textual Similarity +1

Leveraging pre-trained representations to improve access to untranscribed speech from endangered languages

1 code implementation26 Mar 2021 Nay San, Martijn Bartelds, Mitchell Browne, Lily Clifford, Fiona Gibson, John Mansfield, David Nash, Jane Simpson, Myfany Turpin, Maria Vollmer, Sasha Wilmoth, Dan Jurafsky

Surprisingly, the English model outperformed the multilingual model on 4 Australian language datasets, raising questions around how to optimally leverage self-supervised speech representations for QbE-STD.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +1

Detecting Stance in Media on Global Warming

1 code implementation Findings of the Association for Computational Linguistics 2020 Yiwei Luo, Dallas Card, Dan Jurafsky

We release our stance dataset, model, and lexicons of framing devices for future work on opinion-framing and the automatic detection of GW stance.

Causal Effects of Linguistic Properties

1 code implementation NAACL 2021 Reid Pryzant, Dallas Card, Dan Jurafsky, Victor Veitch, Dhanya Sridhar

Second, in practice, we only have access to noisy proxies for the linguistic properties of interest -- e. g., predictions from classifiers and lexicons.

Language Modelling

Improving Factual Completeness and Consistency of Image-to-Text Radiology Report Generation

3 code implementations NAACL 2021 Yasuhide Miura, Yuhao Zhang, Emily Bao Tsai, Curtis P. Langlotz, Dan Jurafsky

We further show via a human evaluation and a qualitative analysis that our system leads to generations that are more factually complete and consistent compared to the baselines.

Image to text Natural Language Inference +1

With Little Power Comes Great Responsibility

2 code implementations EMNLP 2020 Dallas Card, Peter Henderson, Urvashi Khandelwal, Robin Jia, Kyle Mahowald, Dan Jurafsky

Despite its importance to experimental design, statistical power (the probability that, given a real effect, an experiment will reject the null hypothesis) has largely been ignored by the NLP community.

Experimental Design Machine Translation +1

Nearest Neighbor Machine Translation

5 code implementations ICLR 2021 Urvashi Khandelwal, Angela Fan, Dan Jurafsky, Luke Zettlemoyer, Mike Lewis

We introduce $k$-nearest-neighbor machine translation ($k$NN-MT), which predicts tokens with a nearest neighbor classifier over a large datastore of cached examples, using representations from a neural translation model for similarity search.

Decoder Machine Translation +1

Utility is in the Eye of the User: A Critique of NLP Leaderboards

no code implementations EMNLP 2020 Kawin Ethayarajh, Dan Jurafsky

Benchmarks such as GLUE have helped drive advances in NLP by incentivizing the creation of more accurate models.

Fairness

The Role of Verb Semantics in Hungarian Verb-Object Order

no code implementations16 Jun 2020 Dorottya Demszky, László Kálmán, Dan Jurafsky, Beth Levin

We test the effect of lexical semantics on the ordering of verbs and their objects by grouping verbs into 11 semantic classes.

Object

Pretraining with Contrastive Sentence Objectives Improves Discourse Performance of Language Models

1 code implementation ACL 2020 Dan Iter, Kelvin Guu, Larry Lansing, Dan Jurafsky

Recent models for unsupervised representation learning of text have employed a number of techniques to improve contextual word representations but have put little focus on discourse-level representations.

Common Sense Reasoning Natural Language Inference +4

A Framework for the Computational Linguistic Analysis of Dehumanization

no code implementations6 Mar 2020 Julia Mendelsohn, Yulia Tsvetkov, Dan Jurafsky

Dehumanization is a pernicious psychological process that often leads to extreme intergroup bias, hate speech, and violence aimed at targeted social groups.

Abusive Language

Automatically Neutralizing Subjective Bias in Text

1 code implementation21 Nov 2019 Reid Pryzant, Richard Diehl Martinez, Nathan Dass, Sadao Kurohashi, Dan Jurafsky, Diyi Yang

To address this issue, we introduce a novel testbed for natural language generation: automatically bringing inappropriately subjective text into a neutral point of view ("neutralizing" biased text).

Decoder Sentence +1

Social Bias Frames: Reasoning about Social and Power Implications of Language

no code implementations ACL 2020 Maarten Sap, Saadia Gabriel, Lianhui Qin, Dan Jurafsky, Noah A. Smith, Yejin Choi

We introduce Social Bias Frames, a new conceptual formalism that aims to model the pragmatic frames in which people project social biases and stereotypes onto others.

Generalization through Memorization: Nearest Neighbor Language Models

5 code implementations ICLR 2020 Urvashi Khandelwal, Omer Levy, Dan Jurafsky, Luke Zettlemoyer, Mike Lewis

Applying this augmentation to a strong Wikitext-103 LM, with neighbors drawn from the original training set, our $k$NN-LM achieves a new state-of-the-art perplexity of 15. 79 - a 2. 9 point improvement with no additional training.

Domain Adaptation Language Modelling +1

From Insanely Jealous to Insanely Delicious: Computational Models for the Semantic Bleaching of English Intensifiers

no code implementations WS 2019 Yiwei Luo, Dan Jurafsky, Beth Levin

We introduce novel computational models for modeling semantic bleaching, a widespread category of change in which words become more abstract or lose elements of meaning, like the development of {``}arrive{''} from its earlier meaning {`}become at shore.

Neural Text Style Transfer via Denoising and Reranking

no code implementations WS 2019 Joseph Lee, Ziang Xie, Cindy Wang, Max Drach, Dan Jurafsky, Andrew Ng

We introduce a simple method for text style transfer that frames style transfer as denoising: we synthesize a noisy corpus and treat the source style as a noisy version of the target style.

Denoising Style Transfer +1

Let's Make Your Request More Persuasive: Modeling Persuasive Strategies via Semi-Supervised Neural Nets on Crowdfunding Platforms

no code implementations NAACL 2019 Diyi Yang, Jiaao Chen, Zichao Yang, Dan Jurafsky, Eduard Hovy

Modeling what makes a request persuasive - eliciting the desired response from a reader - is critical to the study of propaganda, behavioral economics, and advertising.

Persuasiveness Sentence

Sample Efficient Text Summarization Using a Single Pre-Trained Transformer

2 code implementations21 May 2019 Urvashi Khandelwal, Kevin Clark, Dan Jurafsky, Lukasz Kaiser

Language model (LM) pre-training has resulted in impressive performance and sample efficiency on a variety of language understanding tasks.

 Ranked #1 on Text Summarization on DUC 2004 Task 1 (ROUGE-2 metric)

Abstractive Text Summarization Decoder +1

Integrating Text and Image: Determining Multimodal Document Intent in Instagram Posts

1 code implementation IJCNLP 2019 Julia Kruk, Jonah Lubin, Karan Sikka, Xiao Lin, Dan Jurafsky, Ajay Divakaran

Computing author intent from multimodal data like Instagram posts requires modeling a complex relationship between text and image.

Intent Detection

Analyzing Polarization in Social Media: Method and Application to Tweets on 21 Mass Shootings

1 code implementation NAACL 2019 Dorottya Demszky, Nikhil Garg, Rob Voigt, James Zou, Matthew Gentzkow, Jesse Shapiro, Dan Jurafsky

We provide an NLP framework to uncover four linguistic dimensions of political polarization in social media: topic choice, framing, affect and illocutionary force.

Clustering

Textual Analogy Parsing: What's Shared and What's Compared among Analogous Facts

2 code implementations EMNLP 2018 Matthew Lamm, Arun Tejasvi Chaganty, Christopher D. Manning, Dan Jurafsky, Percy Liang

To understand a sentence like "whereas only 10% of White Americans live at or below the poverty line, 28% of African Americans do" it is important not only to identify individual facts, e. g., poverty rates of distinct demographic groups, but also the higher-order relations between them, e. g., the disparity between them.

Sentence Textual Analogy Parsing

Embedding Logical Queries on Knowledge Graphs

6 code implementations NeurIPS 2018 William L. Hamilton, Payal Bajaj, Marinka Zitnik, Dan Jurafsky, Jure Leskovec

Learning low-dimensional embeddings of knowledge graphs is a powerful approach used to predict unobserved or missing edges between entities.

Complex Query Answering

Automatic Detection of Incoherent Speech for Diagnosing Schizophrenia

no code implementations WS 2018 Dan Iter, Jong Yoon, Dan Jurafsky

Here, we present the first benchmark comparison of previously proposed coherence models for detecting symptoms of schizophrenia and evaluate their performance on a new dataset of recorded interviews between subjects and clinicians.

Sentence Sentence Embedding +2

Noising and Denoising Natural Language: Diverse Backtranslation for Grammar Correction

no code implementations NAACL 2018 Ziang Xie, Guillaume Genthial, Stanley Xie, Andrew Ng, Dan Jurafsky

Translation-based methods for grammar correction that directly map noisy, ungrammatical text to their clean counterparts are able to correct a broad range of errors; however, such techniques are bottlenecked by the need for a large parallel corpus of noisy and clean sentence pairs.

Denoising Machine Translation +2

Deconfounded Lexicon Induction for Interpretable Social Science

no code implementations NAACL 2018 Reid Pryzant, Kelly Shen, Dan Jurafsky, Stefan Wagner

The first uses a bifurcated architecture to separate the explanatory power of the text and confounds.

Community Interaction and Conflict on the Web

no code implementations9 Mar 2018 Srijan Kumar, William L. Hamilton, Jure Leskovec, Dan Jurafsky

Here we study intercommunity interactions across 36, 000 communities on Reddit, examining cases where users of one community are mobilized by negative sentiment to comment in another community.

Detecting Institutional Dialog Acts in Police Traffic Stops

no code implementations TACL 2018 Vinodkumar Prabhakaran, Camilla Griffiths, Hang Su, Prateek Verma, Nelson Morgan, Jennifer L. Eberhardt, Dan Jurafsky

We apply computational dialog methods to police body-worn camera footage to model conversations between police officers and community members in traffic stops.

speech-recognition Speech Recognition

Word Embeddings Quantify 100 Years of Gender and Ethnic Stereotypes

1 code implementation22 Nov 2017 Nikhil Garg, Londa Schiebinger, Dan Jurafsky, James Zou

Word embeddings use vectors to represent words such that the geometry between vectors captures semantic relationship between the words.

Word Embeddings

Neural Net Models of Open-domain Discourse Coherence

no code implementations EMNLP 2017 Jiwei Li, Dan Jurafsky

In this paper, we describe domain-independent neural models of discourse coherence that are capable of measuring multiple aspects of coherence in existing sentences and can maintain coherence while generating new sentences.

Abstractive Text Summarization Question Answering +2

Community Identity and User Engagement in a Multi-Community Landscape

no code implementations26 May 2017 Justine Zhang, William L. Hamilton, Cristian Danescu-Niculescu-Mizil, Dan Jurafsky, Jure Leskovec

To this end we introduce a quantitative, language-based typology reflecting two key aspects of a community's identity: how distinctive, and how temporally dynamic it is.

A Two-stage Sieve Approach for Quote Attribution

no code implementations EACL 2017 Grace Muzny, Michael Fang, Angel Chang, Dan Jurafsky

We present a deterministic sieve-based system for attributing quotations in literary text and a new dataset: QuoteLi3.

Vocal Bursts Valence Prediction

Loyalty in Online Communities

1 code implementation9 Mar 2017 William L. Hamilton, Justine Zhang, Cristian Danescu-Niculescu-Mizil, Dan Jurafsky, Jure Leskovec

In this paper we operationalize loyalty as a user-community relation: users loyal to a community consistently prefer it over all others; loyal communities retain their loyal users over time.

Data Distillation for Controlling Specificity in Dialogue Generation

no code implementations22 Feb 2017 Jiwei Li, Will Monroe, Dan Jurafsky

We show that from such a set of subsystems, one can use reinforcement learning to build a system that tailors its output to different input contexts at test time.

Dialogue Generation reinforcement-learning +3

Adversarial Learning for Neural Dialogue Generation

8 code implementations EMNLP 2017 Jiwei Li, Will Monroe, Tianlin Shi, Sébastien Jean, Alan Ritter, Dan Jurafsky

In this paper, drawing intuition from the Turing test, we propose using adversarial training for open-domain dialogue generation: the system is trained to produce sequences that are indistinguishable from human-generated dialogue utterances.

Dialogue Evaluation Dialogue Generation +2

Learning to Decode for Future Success

no code implementations23 Jan 2017 Jiwei Li, Will Monroe, Dan Jurafsky

We introduce a simple, general strategy to manipulate the behavior of a neural decoder that enables it to generate outputs that have specific properties of interest (e. g., sequences of a pre-specified length).

Abstractive Text Summarization Decision Making +3

Understanding Neural Networks through Representation Erasure

no code implementations24 Dec 2016 Jiwei Li, Will Monroe, Dan Jurafsky

While neural networks have been successfully applied to many natural language processing tasks, they come at the cost of interpretability.

Reinforcement Learning Sentence +1

A Simple, Fast Diverse Decoding Algorithm for Neural Generation

1 code implementation25 Nov 2016 Jiwei Li, Will Monroe, Dan Jurafsky

We further propose a variation that is capable of automatically adjusting its diversity decoding rates for different inputs using reinforcement learning (RL).

Abstractive Text Summarization Diversity +5

Citation Classification for Behavioral Analysis of a Scientific Field

no code implementations2 Sep 2016 David Jurgens, Srijan Kumar, Raine Hoover, Dan McFarland, Dan Jurafsky

Citations are an important indicator of the state of a scientific field, reflecting how authors frame their work, and influencing uptake by future scholars.

Classification General Classification

Cultural Shift or Linguistic Drift? Comparing Two Computational Measures of Semantic Change

no code implementations EMNLP 2016 William L. Hamilton, Jure Leskovec, Dan Jurafsky

Words shift in meaning for many reasons, including cultural factors like new technologies and regular linguistic processes like subjectification.

Cultural Vocal Bursts Intensity Prediction

Deep Reinforcement Learning for Dialogue Generation

8 code implementations EMNLP 2016 Jiwei Li, Will Monroe, Alan Ritter, Michel Galley, Jianfeng Gao, Dan Jurafsky

Recent neural models of dialogue generation offer great promise for generating responses for conversational agents, but tend to be shortsighted, predicting utterances one at a time while ignoring their influence on future outcomes.

Deep Reinforcement Learning Dialogue Generation +4

Neural Net Models for Open-Domain Discourse Coherence

1 code implementation5 Jun 2016 Jiwei Li, Dan Jurafsky

In this paper, we describe domain-independent neural models of discourse coherence that are capable of measuring multiple aspects of coherence in existing sentences and can maintain coherence while generating new sentences.

Text Generation

Diachronic Word Embeddings Reveal Statistical Laws of Semantic Change

6 code implementations ACL 2016 William L. Hamilton, Jure Leskovec, Dan Jurafsky

Understanding how words change their meanings over time is key to models of language and cultural evolution, but historical data on meaning is scarce, making theories hard to develop and test.

Diachronic Word Embeddings Word Embeddings

Neural Language Correction with Character-Based Attention

3 code implementations31 Mar 2016 Ziang Xie, Anand Avati, Naveen Arivazhagan, Dan Jurafsky, Andrew Y. Ng

Motivated by these issues, we present a neural network-based approach to language correction.

Decoder Language Modelling +2

Mutual Information and Diverse Decoding Improve Neural Machine Translation

1 code implementation4 Jan 2016 Jiwei Li, Dan Jurafsky

We introduce an alternative objective function for neural MT that maximizes the mutual information between the source and target sentences, modeling the bi-directional dependency of sources and targets.

Diversity Machine Translation +3

Learning multi-faceted representations of individuals from heterogeneous evidence using neural networks

no code implementations18 Oct 2015 Jiwei Li, Alan Ritter, Dan Jurafsky

Inferring latent attributes of people online is an important social computing task, but requires integrating the many heterogeneous sources of information available on the web.

Community Detection Link Prediction

Visualizing and Understanding Neural Models in NLP

1 code implementation NAACL 2016 Jiwei Li, Xinlei Chen, Eduard Hovy, Dan Jurafsky

While neural networks have been successfully applied to many NLP tasks the resulting vector-based models are very difficult to interpret.

Negation Sentence

Do Multi-Sense Embeddings Improve Natural Language Understanding?

no code implementations EMNLP 2015 Jiwei Li, Dan Jurafsky

Learning a distinct representation for each sense of an ambiguous word could lead to more powerful and fine-grained models of vector-space representations.

named-entity-recognition Named Entity Recognition +6

A Hierarchical Neural Autoencoder for Paragraphs and Documents

6 code implementations IJCNLP 2015 Jiwei Li, Minh-Thang Luong, Dan Jurafsky

Natural language generation of coherent long texts like paragraphs or longer documents is a challenging problem for recurrent networks models.

Sentence Text Generation

When Are Tree Structures Necessary for Deep Learning of Representations?

no code implementations EMNLP 2015 Jiwei Li, Minh-Thang Luong, Dan Jurafsky, Eudard Hovy

Recursive neural models, which use syntactic parse trees to recursively generate representations bottom-up, are a popular architecture.

Deep Learning Discourse Parsing +5

Inferring User Preferences by Probabilistic Logical Reasoning over Social Networks

no code implementations11 Nov 2014 Jiwei Li, Alan Ritter, Dan Jurafsky

by building a probabilistic model that reasons over user attributes (the user's location or gender) and the social network (the user's friends and spouse), via inferences like homophily (I am more likely to like sushi if spouse or friends like sushi, I am more likely to like the Knicks if I live in New York).

Attribute Logical Reasoning +1

How to Ask for a Favor: A Case Study on the Success of Altruistic Requests

no code implementations13 May 2014 Tim Althoff, Cristian Danescu-Niculescu-Mizil, Dan Jurafsky

We present a case study of altruistic requests in an online community where all requests ask for the very same contribution and do not offer anything tangible in return, allowing us to disentangle what is requested from textual and social factors.

Cannot find the paper you are looking for? You can Submit a new open access paper.