Search Results for author: Jianbin Fang

Found 4 papers, 1 papers with code

Optimizing Streaming Parallelism on Heterogeneous Many-Core Architectures: A Machine Learning Based Approach

1 code implementation5 Mar 2020 Peng Zhang, Jianbin Fang, Canqun Yang, Chun Huang, Tao Tang, Zheng Wang

This article presents an automatic approach to quickly derive a good solution for hardware resource partition and task granularity for task-based parallel applications on heterogeneous many-core architectures.

BIG-bench Machine Learning

Characterizing Scalability of Sparse Matrix-Vector Multiplications on Phytium FT-2000+ Many-cores

no code implementations20 Nov 2019 Donglin Chen, Jianbin Fang, Chuanfu Xu, Shizhao Chen, Zheng Wang

Understanding the scalability of parallel programs is crucial for software optimization and hardware architecture design.

To Compress, or Not to Compress: Characterizing Deep Learning Model Compression for Embedded Inference

no code implementations21 Oct 2018 Qing Qin, Jie Ren, Jialong Yu, Ling Gao, Hai Wang, Jie Zheng, Yansong Feng, Jianbin Fang, Zheng Wang

We experimentally show that how two mainstream compression techniques, data quantization and pruning, perform on these network architectures and the implications of compression techniques to the model storage size, inference time, energy consumption and performance metrics.

Image Classification Model Compression +1

Tuning Streamed Applications on Intel Xeon Phi: A Machine Learning Based Approach

no code implementations8 Feb 2018 Peng Zhang, Jianbin Fang, Tao Tang, Canqun Yang, Zheng Wang

In this paper, we present an automatic approach to determine the hardware resource partition and the task granularity for any given application, targeting the Intel XeonPhi architecture.

Performance

Cannot find the paper you are looking for? You can Submit a new open access paper.