Search Results for author: Gail Weiss

Found 9 papers, 4 papers with code

Transformers as Recognizers of Formal Languages: A Survey on Expressivity

no code implementations1 Nov 2023 Lena Strobl, William Merrill, Gail Weiss, David Chiang, Dana Angluin

As transformers have gained prominence in natural language processing, some researchers have investigated theoretically what problems they can and cannot solve, by treating problems as formal languages.

Discovering Knowledge-Critical Subnetworks in Pretrained Language Models

no code implementations4 Oct 2023 Deniz Bayazit, Negar Foroutan, Zeming Chen, Gail Weiss, Antoine Bosselut

In this work, we investigate whether pretrained language models contain various knowledge-critical subnetworks: particular sparse computational subgraphs responsible for encoding specific knowledge the model has memorized.

Language Modelling

RECKONING: Reasoning through Dynamic Knowledge Encoding

no code implementations NeurIPS 2023 Zeming Chen, Gail Weiss, Eric Mitchell, Asli Celikyilmaz, Antoine Bosselut

In the outer loop, the model learns to use the updated weights to reproduce and answer reasoning questions about the memorized knowledge.

Thinking Like Transformers

4 code implementations13 Jun 2021 Gail Weiss, Yoav Goldberg, Eran Yahav

In this paper we aim to change that, proposing a computational model for the transformer-encoder in the form of a programming language.

Synthesizing Context-free Grammars from Recurrent Neural Networks (Extended Version)

no code implementations20 Jan 2021 Daniel M. Yellin, Gail Weiss

We present an algorithm for extracting a subclass of the context free grammars (CFGs) from a trained recurrent neural network (RNN).

A Formal Hierarchy of RNN Architectures

no code implementations ACL 2020 William Merrill, Gail Weiss, Yoav Goldberg, Roy Schwartz, Noah A. Smith, Eran Yahav

While formally extending these findings to unsaturated RNNs is left to future work, we hypothesize that the practical learnable capacity of unsaturated RNNs obeys a similar hierarchy.

Learning Deterministic Weighted Automata with Queries and Counterexamples

1 code implementation NeurIPS 2019 Gail Weiss, Yoav Goldberg, Eran Yahav

We present an algorithm for extraction of a probabilistic deterministic finite automaton (PDFA) from a given black-box language model, such as a recurrent neural network (RNN).

Language Modelling

On the Practical Computational Power of Finite Precision RNNs for Language Recognition

1 code implementation ACL 2018 Gail Weiss, Yoav Goldberg, Eran Yahav

While Recurrent Neural Networks (RNNs) are famously known to be Turing complete, this relies on infinite precision in the states and unbounded computation time.

Cannot find the paper you are looking for? You can Submit a new open access paper.