1 code implementation • 3 Mar 2024 • Ameen Ali, Itamar Zimerman, Lior Wolf
The Mamba layer offers an efficient selective state space model (SSM) that is highly effective in modeling multiple domains, including NLP, long-range sequence processing, and computer vision.
no code implementations • 28 Nov 2023 • Itamar Zimerman, Lior Wolf
Despite their dominance in modern DL and, especially, NLP domains, transformer architectures exhibit sub-optimal performance on long-range tasks compared to recent layers that are specifically designed for this purpose.
no code implementations • 15 Nov 2023 • Itamar Zimerman, Moran Baruch, Nir Drucker, Gilad Ezov, Omri Soceanu, Lior Wolf
This innovation enables us to perform secure inference on LMs with WikiText-103.
no code implementations • 24 Sep 2023 • Itamar Zimerman, Lior Wolf
Our empirical findings indicate that the proposed Hyena N-D layer boosts the performance of various Vision Transformer architectures, such as ViT, Swin, and DeiT across multiple datasets.
no code implementations • 11 Jun 2023 • Nir Drucker, Itamar Zimerman
Homomorphic Encryption (HE) is a cryptographic tool that allows performing computation under encryption, which is used by many privacy-preserving machine learning solutions, for example, to perform secure classification.
1 code implementation • 11 Jun 2023 • Ethan Baron, Itamar Zimerman, Lior Wolf
For example, vision transformers equipped with our layer exhibit effective performance even without positional encoding
no code implementations • 8 Jun 2023 • Shmuel Bar-David, Itamar Zimerman, Eliya Nachmani, Lior Wolf
Recently, sequence learning methods have been applied to the problem of off-policy Reinforcement Learning, including the seminal work on Decision Transformers, which employs transformers for this task.
no code implementations • 24 May 2023 • Shahar Lutati, Itamar Zimerman, Lior Wolf
We present a new layer in which dynamic (i. e., input-dependent) Infinite Impulse Response (IIR) filters of order two are used to process the input sequence prior to applying conventional attention.
Ranked #1 on Language Modelling on Text8
no code implementations • 26 Apr 2023 • Moran Baruch, Nir Drucker, Gilad Ezov, Yoav Goldberg, Eyal Kushnir, Jenny Lerner, Omri Soceanu, Itamar Zimerman
Training large-scale CNNs that during inference can be run under Homomorphic Encryption (HE) is challenging due to the need to use only polynomial operations.
no code implementations • 9 Jun 2021 • Itamar Zimerman, Eliya Nachmani, Lior Wolf
In this work, we combine a novel cryptographic variant of a deep error correcting code technique with a modified SAT solver scheme to apply the attack on AES keys.