no code implementations • 20 Feb 2020 • Valentin Radu, Kuba Kaszyk, Yuan Wen, Jack Turner, Jose Cano, Elliot J. Crowley, Bjorn Franke, Amos Storkey, Michael O'Boyle
We evaluate higher level libraries, which analyze the input characteristics of a convolutional layer, based on which they produce optimized OpenCL (Arm Compute Library and TVM) and CUDA (cuDNN) code.
no code implementations • 21 May 2020 • Yuan Wen, Andrew Anderson, Valentin Radu, Michael F. P. O'Boyle, David Gregg
We optimize the trade-off between execution time and memory consumption by: 1) attempting to minimize execution time across the whole network by selecting data layouts and primitive operations to implement each layer; and 2) allocating an appropriate workspace that reflects the upper bound of memory footprint per layer.
no code implementations • 22 Jun 2020 • Yuan Wen, David Gregg
Pruning and quantization are proven methods for improving the performance and storage efficiency of convolutional neural networks (CNNs).
no code implementations • LILT 2019 • Bin Li, Yuan Wen, Li Song, Weiguang Qu, Nianwen Xue
One significant change we have made to the AMR annotation methodology is the inclusion of the alignment between word tokens in the sentence and the concepts/relations in the CAMR annotation to make it easier for automatic parsers to model the correspondence between a sentence and its meaning representation.