Search Results for author: Lei Xun

Found 6 papers, 3 papers with code

Dynamic DNNs and Runtime Management for Efficient Inference on Mobile/Embedded Devices

1 code implementation17 Jan 2024 Lei Xun, Jonathon Hare, Geoff V. Merrett

In this thesis, we proposed a combined method, a system was developed for DNN performance trade-off management, combining the runtime trade-off opportunities in both algorithms and hardware to meet dynamically changing application performance targets and hardware constraints in real time.

Management Model Compression

Dynamic Transformer for Efficient Machine Translation on Embedded Devices

no code implementations17 Jul 2021 Hishan Parry, Lei Xun, Amin Sabet, Jia Bi, Jonathon Hare, Geoff V. Merrett

The new reduced design space results in a BLEU score increase of approximately 1% for sub-optimal models from the original design space, with a wide range for performance scaling between 0. 356s - 1. 526s for the GPU and 2. 9s - 7. 31s for the CPU.

Machine Translation Translation

Optimising Resource Management for Embedded Machine Learning

1 code implementation8 May 2021 Lei Xun, Long Tran-Thanh, Bashir M Al-Hashimi, Geoff V. Merrett

Machine learning inference is increasingly being executed locally on mobile and embedded platforms, due to the clear advantages in latency, privacy and connectivity.

BIG-bench Machine Learning Management

Incremental Training and Group Convolution Pruning for Runtime DNN Performance Scaling on Heterogeneous Embedded Platforms

no code implementations8 May 2021 Lei Xun, Long Tran-Thanh, Bashir M Al-Hashimi, Geoff V. Merrett

Compared to the existing works, our approach can provide up to 2. 36x (energy) and 2. 73x (time) wider dynamic range with a 2. 4x smaller memory footprint at the same compression rate.

Dynamic-OFA: Runtime DNN Architecture Switching for Performance Scaling on Heterogeneous Embedded Platforms

1 code implementation8 May 2021 Wei Lou, Lei Xun, Amin Sabet, Jia Bi, Jonathon Hare, Geoff V. Merrett

However, the training process of such dynamic DNNs can be costly, since platform-aware models of different deployment scenarios must be retrained to become dynamic.

Cannot find the paper you are looking for? You can Submit a new open access paper.