Search Results for author: Michael Hanna

Found 9 papers, 5 papers with code

Analyzing BERT’s Knowledge of Hypernymy via Prompting

no code implementations EMNLP (BlackboxNLP) 2021 Michael Hanna, David Mareček

The high performance of large pretrained language models (LLMs) such as BERT on NLP tasks has prompted questions about BERT’s linguistic capabilities, and how they differ from humans’.

Hypernym Discovery

A Fine-Grained Analysis of BERTScore

no code implementations WMT (EMNLP) 2021 Michael Hanna, Ondřej Bojar

BERTScore, a recently proposed automatic metric for machine translation quality, uses BERT, a large pre-trained language model to evaluate candidate translations with respect to a gold translation.

Language Modelling Machine Translation +4

Have Faith in Faithfulness: Going Beyond Circuit Overlap When Finding Model Mechanisms

no code implementations26 Mar 2024 Michael Hanna, Sandro Pezzelle, Yonatan Belinkov

Most studies determine which edges belong in a LM's circuit by performing causal interventions on each edge independently, but this scales poorly with model size.

Language Modelling

Do Pre-Trained Language Models Detect and Understand Semantic Underspecification? Ask the DUST!

1 code implementation19 Feb 2024 Frank Wildenburg, Michael Hanna, Sandro Pezzelle

In this work, we propose a novel Dataset of semantically Underspecified Sentences grouped by Type (DUST) and use it to study whether pre-trained language models (LMs) correctly identify and interpret underspecified sentences.

Sentence

When Language Models Fall in Love: Animacy Processing in Transformer Language Models

1 code implementation23 Oct 2023 Michael Hanna, Yonatan Belinkov, Sandro Pezzelle

However, we also show that even when presented with stories about atypically animate entities, such as a peanut in love, LMs adapt: they treat these entities as animate, though they do not adapt as well as humans.

Cannot find the paper you are looking for? You can Submit a new open access paper.