Search Results for author: Guy Dar

Found 5 papers, 5 papers with code

Memory-efficient Transformers via Top-k Attention

1 code implementation EMNLP (sustainlp) 2021 Ankit Gupta, Guy Dar, Shaya Goodman, David Ciprut, Jonathan Berant

Following the success of dot-product attention in Transformers, numerous approximations have been recently proposed to address its quadratic complexity with respect to the input length.

In-context Learning and Gradient Descent Revisited

1 code implementation13 Nov 2023 Gilad Deutch, Nadav Magar, Tomer Bar Natan, Guy Dar

Next, we explore a major discrepancy in the flow of information throughout the model between ICL and GD, which we term Layer Causality.

Few-Shot Learning In-Context Learning

Analyzing Transformers in Embedding Space

1 code implementation6 Sep 2022 Guy Dar, Mor Geva, Ankit Gupta, Jonathan Berant

In this work, we present a theoretical analysis where all parameters of a trained Transformer are interpreted by projecting them into the embedding space, that is, the space of vocabulary items they operate on.

LM-Debugger: An Interactive Tool for Inspection and Intervention in Transformer-Based Language Models

1 code implementation26 Apr 2022 Mor Geva, Avi Caciularu, Guy Dar, Paul Roit, Shoval Sadde, Micah Shlain, Bar Tamir, Yoav Goldberg

The opaque nature and unexplained behavior of transformer-based language models (LMs) have spurred a wide interest in interpreting their predictions.

Memory-efficient Transformers via Top-$k$ Attention

1 code implementation13 Jun 2021 Ankit Gupta, Guy Dar, Shaya Goodman, David Ciprut, Jonathan Berant

Following the success of dot-product attention in Transformers, numerous approximations have been recently proposed to address its quadratic complexity with respect to the input length.

Cannot find the paper you are looking for? You can Submit a new open access paper.