no code implementations • 22 Jun 2017 • Shuchang Zhou, Yuzhi Wang, He Wen, Qinyao He, Yuheng Zou
Overall, our method improves the prediction accuracies of QNNs without introducing extra computation during inference, has negligible impact on training speed, and is applicable to both Convolutional Neural Networks and Recurrent Neural Networks.
2 code implementations • 30 Nov 2016 • Qinyao He, He Wen, Shuchang Zhou, Yuxin Wu, Cong Yao, Xinyu Zhou, Yuheng Zou
In addition, we propose balanced quantization methods for weights to further reduce performance degradation.
12 code implementations • 20 Jun 2016 • Shuchang Zhou, Yuxin Wu, Zekun Ni, Xinyu Zhou, He Wen, Yuheng Zou
We propose DoReFa-Net, a method to train convolutional neural networks that have low bitwidth weights and activations using low bitwidth parameter gradients.