Search Results for author: Emir Haleva

Found 2 papers, 2 papers with code

Linear Log-Normal Attention with Unbiased Concentration

1 code implementation22 Nov 2023 Yury Nahshan, Joseph Kampeas, Emir Haleva

Transformer models have achieved remarkable results in a wide range of applications.

Rotation Invariant Quantization for Model Compression

1 code implementation3 Mar 2023 Joseph Kampeas, Yury Nahshan, Hanoch Kremer, Gil Lederman, Shira Zaloshinski, Zheng Li, Emir Haleva

Post-training Neural Network (NN) model compression is an attractive approach for deploying large, memory-consuming models on devices with limited memory resources.

Model Compression Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.