Search Results for author: Helen Ngo

Found 5 papers, 1 papers with code

No News is Good News: A Critique of the One Billion Word Benchmark

no code implementations25 Oct 2021 Helen Ngo, João G. M. Araújo, Jeffrey Hui, Nicholas Frosst

The One Billion Word Benchmark is a dataset derived from the WMT 2011 News Crawl, commonly used to measure language modeling ability in natural language processing.

Language Modelling

Mitigating harm in language models with conditional-likelihood filtration

no code implementations4 Aug 2021 Helen Ngo, Cooper Raterink, João G. M. Araújo, Ivan Zhang, Carol Chen, Adrien Morisot, Nicholas Frosst

Language models trained on large-scale unfiltered datasets curated from the open web acquire systemic biases, prejudices, and harmful views from their training data.

Language Modelling

Cannot find the paper you are looking for? You can Submit a new open access paper.