Search Results for author: Michael Hahn

Found 15 papers, 6 papers with code

An Information-Theoretic Characterization of Morphological Fusion

1 code implementation EMNLP 2021 Neil Rathi, Michael Hahn, Richard Futrell

Linguistic typology generally divides synthetic languages into groups based on their morphological fusion.

Modeling Task Effects in Human Reading with Neural Network-based Attention

no code implementations31 Jul 2018 Michael Hahn, Frank Keller

Research on human reading has long documented that reading behavior shows task-specific effects, but it has been challenging to build general models predicting what reading behavior humans will show in a given task.

Question Answering Reading Comprehension

Character-based Surprisal as a Model of Reading Difficulty in the Presence of Error

no code implementations2 Feb 2019 Michael Hahn, Frank Keller, Yonatan Bisk, Yonatan Belinkov

Also, transpositions are more difficult than misspellings, and a high error rate increases difficulty for all words, including correct ones.

Theoretical Limitations of Self-Attention in Neural Sequence Models

no code implementations TACL 2020 Michael Hahn

These limitations seem surprising given the practical success of self-attention and the prominent role assigned to hierarchical structure in linguistics, suggesting that natural language can be approximated well with models that are too weak for the formal languages typically assumed in theoretical linguistics.

Hard Attention

Sensitivity as a Complexity Measure for Sequence Classification Tasks

1 code implementation21 Apr 2021 Michael Hahn, Dan Jurafsky, Richard Futrell

We introduce a theoretical framework for understanding and predicting the complexity of sequence classification tasks, using a novel extension of the theory of Boolean function sensitivity.

General Classification text-classification +1

Crosslinguistic word order variation reflects evolutionary pressures of dependency and information locality

1 code implementation9 Jun 2022 Michael Hahn, Yang Xu

Using data from 80 languages in 17 language families and phylogenetic modeling, we demonstrate that languages evolve to balance these pressures, such that word order change is accompanied by change in the frequency distribution of the syntactic structures which speakers communicate to maintain overall efficiency.

A Theory of Emergent In-Context Learning as Implicit Structure Induction

no code implementations14 Mar 2023 Michael Hahn, Navin Goyal

Scaling large language models (LLMs) leads to an emergent capacity to learn in-context from example demonstrations.

In-Context Learning

Why are Sensitive Functions Hard for Transformers?

no code implementations15 Feb 2024 Michael Hahn, Mark Rofin

We show theoretically and empirically that this theory unifies a broad array of empirical observations about the learning abilities and biases of transformers, such as their generalization bias towards low sensitivity and low degree, and difficulty in length generalization for PARITY.

Cannot find the paper you are looking for? You can Submit a new open access paper.