1 code implementation • 2 Oct 2024 • Edan Kinderman, Itay Hubara, Haggai Maron, Daniel Soudry
Many recent methods aim to merge neural networks (NNs) with identical architectures trained on different tasks to obtain a single multi-task model.
no code implementations • 25 Jan 2024 • Yaniv Blumenfeld, Itay Hubara, Daniel Soudry
The majority of the research on the quantization of Deep Neural Networks (DNNs) is focused on reducing the precision of tensors visible by high-level frameworks (e. g., weights, activations, and gradients).
no code implementations • 21 Mar 2022 • Brian Chmiel, Itay Hubara, Ron Banner, Daniel Soudry
We show that while minimization of the MSE works fine for pruning the weights and activations, it catastrophically fails for the neural gradients.
1 code implementation • NeurIPS 2021 • Itay Hubara, Brian Chmiel, Moshe Island, Ron Banner, Seffi Naor, Daniel Soudry
Finally, to solve the problem of switching between different structure constraints, we suggest a method to convert a pre-trained model with unstructured sparsity to an N:M fine-grained block sparsity model with little to no training.
2 code implementations • 1 Jan 2021 • Elad Hoffer, Berry Weinstein, Itay Hubara, Tal Ben-Nun, Torsten Hoefler, Daniel Soudry
Although trained on images of a specific size, it is well established that CNNs can be used to evaluate a wide range of image sizes at test time, by adjusting the size of intermediate feature maps.
1 code implementation • 14 Jun 2020 • Itay Hubara, Yury Nahshan, Yair Hanani, Ron Banner, Daniel Soudry
Instead, these methods only use the calibration set to set the activations' dynamic ranges.
no code implementations • 29 Dec 2019 • Tzofnat Greenberg Toledo, Ben Perach, Itay Hubara, Daniel Soudry, Shahar Kvatinsky
A recent example is the GXNOR framework for stochastic training of ternary (TNN) and binary (BNN) neural networks.
no code implementations • CVPR 2020 • Matan Haroush, Itay Hubara, Elad Hoffer, Daniel Soudry
Then, we demonstrate how these samples can be used to calibrate and fine-tune quantized models without using any real data in the process.
4 code implementations • 6 Nov 2019 • Vijay Janapa Reddi, Christine Cheng, David Kanter, Peter Mattson, Guenther Schmuelling, Carole-Jean Wu, Brian Anderson, Maximilien Breughe, Mark Charlebois, William Chou, Ramesh Chukka, Cody Coleman, Sam Davis, Pan Deng, Greg Diamos, Jared Duke, Dave Fick, J. Scott Gardner, Itay Hubara, Sachin Idgunji, Thomas B. Jablin, Jeff Jiao, Tom St. John, Pankaj Kanwar, David Lee, Jeffery Liao, Anton Lokhmotov, Francisco Massa, Peng Meng, Paulius Micikevicius, Colin Osborne, Gennady Pekhimenko, Arun Tejusve Raghunath Rajan, Dilip Sequeira, Ashish Sirasao, Fei Sun, Hanlin Tang, Michael Thomson, Frank Wei, Ephrem Wu, Lingjie Xu, Koichi Yamada, Bing Yu, George Yuan, Aaron Zhong, Peizhao Zhang, Yuchen Zhou
Machine-learning (ML) hardware and software system demand is burgeoning.
2 code implementations • 12 Aug 2019 • Elad Hoffer, Berry Weinstein, Itay Hubara, Tal Ben-Nun, Torsten Hoefler, Daniel Soudry
Although trained on images of aspecific size, it is well established that CNNs can be used to evaluate a wide range of image sizes at test time, by adjusting the size of intermediate feature maps.
1 code implementation • 27 Jan 2019 • Elad Hoffer, Tal Ben-Nun, Itay Hubara, Niv Giladi, Torsten Hoefler, Daniel Soudry
We analyze the effect of batch augmentation on gradient variance and show that it empirically improves convergence for a wide variety of deep neural networks and datasets.
3 code implementations • NeurIPS 2018 • Ron Banner, Itay Hubara, Elad Hoffer, Daniel Soudry
Armed with this knowledge, we quantize the model parameters, activations and layer gradients to 8-bit, leaving at a higher precision only the final step in the computation of the weight gradients.
5 code implementations • ICLR 2018 • Elad Hoffer, Itay Hubara, Daniel Soudry
Neural networks are commonly used as models for classification for a wide variety of tasks.
1 code implementation • NeurIPS 2017 • Elad Hoffer, Itay Hubara, Daniel Soudry
Following this hypothesis we conducted experiments to show empirically that the "generalization gap" stems from the relatively small number of updates rather than the batch size, and can be completely eliminated by adapting the training regime used.
no code implementations • 21 Nov 2016 • Elad Hoffer, Itay Hubara, Nir Ailon
Convolutional networks have marked their place over the last few years as the best performing model for various visual tasks.
1 code implementation • 7 Nov 2016 • Nadav Bhonker, Shai Rozenberg, Itay Hubara
The environment is expandable, allowing for more video games and consoles to be easily added to the environment, while maintaining the same interface as ALE.
Ranked #1 on SNES Games on F-Zero
no code implementations • 2 Oct 2016 • Elad Hoffer, Itay Hubara, Nir Ailon
Convolutional networks have marked their place over the last few years as the best performing model for various visual tasks.
5 code implementations • 22 Sep 2016 • Itay Hubara, Matthieu Courbariaux, Daniel Soudry, Ran El-Yaniv, Yoshua Bengio
Quantized recurrent neural networks were tested over the Penn Treebank dataset, and achieved comparable accuracy as their 32-bit counterparts using only 4-bits.
26 code implementations • 9 Feb 2016 • Matthieu Courbariaux, Itay Hubara, Daniel Soudry, Ran El-Yaniv, Yoshua Bengio
We introduce a method to train Binarized Neural Networks (BNNs) - neural networks with binary weights and activations at run-time.
2 code implementations • NeurIPS 2016 • Itay Hubara, Daniel Soudry, Ran El Yaniv
We introduce a method to train Binarized Neural Networks (BNNs) - neural networks with binary weights and activations at run-time and when computing the parameters' gradient at train-time.
2 code implementations • NeurIPS 2014 • Daniel Soudry, Itay Hubara, Ron Meir
Using online EP and the central limit theorem we find an analytical approximation to the Bayes update of this posterior, as well as the resulting Bayes estimates of the weights and outputs.