Search Results for author: Tyler A. Chang

Found 13 papers, 7 papers with code

Detecting Hallucination and Coverage Errors in Retrieval Augmented Generation for Controversial Topics

no code implementations13 Mar 2024 Tyler A. Chang, Katrin Tomanek, Jessica Hoffmann, Nithum Thain, Erin Van Liemt, Kathleen Meier-Hellstern, Lucas Dixon

We explore a strategy to handle controversial topics in LLM-based chatbots based on Wikipedia's Neutral Point of View (NPOV) principle: acknowledge the absence of a single true answer and surface multiple perspectives.

Hallucination Retrieval +1

A Bit of a Problem: Measurement Disparities in Dataset Sizes Across Languages

no code implementations1 Mar 2024 Catherine Arnett, Tyler A. Chang, Benjamin K. Bergen

We release a tool to obtain byte premiums for any two languages, enabling comparisons of dataset sizes across languages for more equitable multilingual model development and data practices.

When Is Multilinguality a Curse? Language Modeling for 250 High- and Low-Resource Languages

1 code implementation15 Nov 2023 Tyler A. Chang, Catherine Arnett, Zhuowen Tu, Benjamin K. Bergen

However, concrete evidence for the effects of multilinguality on language modeling performance in individual languages remains scarce.

Language Modelling

Structural Priming Demonstrates Abstract Grammatical Representations in Multilingual Language Models

no code implementations15 Nov 2023 James A. Michaelov, Catherine Arnett, Tyler A. Chang, Benjamin K. Bergen

We measure crosslingual structural priming in large language models, comparing model behavior to human experimental results from eight crosslingual experiments covering six languages, and four monolingual structural priming experiments in three non-English languages.

Sentence

Crosslingual Structural Priming and the Pre-Training Dynamics of Bilingual Language Models

no code implementations11 Oct 2023 Catherine Arnett, Tyler A. Chang, James A. Michaelov, Benjamin K. Bergen

Do multilingual language models share abstract grammatical representations across languages, and if so, when do these develop?

Language Modelling

Characterizing Learning Curves During Language Model Pre-Training: Learning, Forgetting, and Stability

1 code implementation29 Aug 2023 Tyler A. Chang, Zhuowen Tu, Benjamin K. Bergen

We quantify the final surprisal, within-run variability, age of acquisition, forgettability, and cross-run variability of learning curves for individual tokens in context.

Language Modelling

Language Model Behavior: A Comprehensive Survey

1 code implementation20 Mar 2023 Tyler A. Chang, Benjamin K. Bergen

Transformer language models have received widespread public attention, yet their generated text is often surprising even to NLP researchers.

Language Modelling Large Language Model +1

The Geometry of Multilingual Language Model Representations

1 code implementation22 May 2022 Tyler A. Chang, Zhuowen Tu, Benjamin K. Bergen

The subspace means differ along language-sensitive axes that are relatively stable throughout middle layers, and these axes encode information such as token vocabularies.

Cross-Lingual Transfer Transfer Learning +1

Word Acquisition in Neural Language Models

1 code implementation5 Oct 2021 Tyler A. Chang, Benjamin K. Bergen

We investigate how neural language models acquire individual words during training, extracting learning curves and ages of acquisition for over 600 words on the MacArthur-Bates Communicative Development Inventory (Fenson et al., 2007).

Language Acquisition

Encodings of Source Syntax: Similarities in NMT Representations Across Target Languages

no code implementations WS 2020 Tyler A. Chang, Anna N. Rafferty

We train neural machine translation (NMT) models from English to six target languages, using NMT encoder representations to predict ancestor constituent labels of source language words.

Machine Translation NMT +1

Cannot find the paper you are looking for? You can Submit a new open access paper.