Search Results for author: Zeye Liu

Found 4 papers, 1 papers with code

FLightNNs: Lightweight Quantized Deep Neural Networks for Fast and Accurate Inference

no code implementations5 Apr 2019 Ruizhou Ding, Zeye Liu, Ting-Wu Chin, Diana Marculescu, R. D., Blanton

Over 46 FPGA-design experiments involving eight configurations and four data sets reveal that lightweight neural networks with a flexible $k$ value (dubbed FLightNNs) fully utilize the hardware resources on Field Programmable Gate Arrays (FPGAs), our experimental results show that FLightNNs can achieve 2$\times$ speedup when compared to lightweight NNs with $k=2$, with only 0. 1\% accuracy degradation.

Quantization

Regularizing Activation Distribution for Training Binarized Deep Networks

1 code implementation CVPR 2019 Ruizhou Ding, Ting-Wu Chin, Zeye Liu, Diana Marculescu

Binarized Neural Networks (BNNs) can significantly reduce the inference latency and energy consumption in resource-constrained devices due to their pure-logical computation and fewer memory accesses.

Differentiable Training for Hardware Efficient LightNNs

no code implementations NIPS Workshop CDNNRIA 2018 Ruizhou Ding, Zeye Liu, Ting-Wu Chin, Diana Marculescu, R.D. (Shawn) Blanton

To reduce runtime and resource utilization of Deep Neural Networks (DNNs) on customized hardware, LightNN has been proposed by constraining the weights of DNNs to be a sum of a limited number (denoted as $k\in\{1, 2\}$) of powers of 2.

Quantization

LightNN: Filling the Gap between Conventional Deep Neural Networks and Binarized Networks

no code implementations2 Dec 2017 Ruizhou Ding, Zeye Liu, Rongye Shi, Diana Marculescu, R. D. Blanton

For a fixed DNN configuration, LightNNs have better accuracy at a slight energy increase than BNNs, yet are more energy efficient with only slightly less accuracy than conventional DNNs.

Cannot find the paper you are looking for? You can Submit a new open access paper.