Search Results for author: Geoff V. Merrett

Found 8 papers, 4 papers with code

Dynamic DNNs and Runtime Management for Efficient Inference on Mobile/Embedded Devices

1 code implementation17 Jan 2024 Lei Xun, Jonathon Hare, Geoff V. Merrett

In this thesis, we proposed a combined method, a system was developed for DNN performance trade-off management, combining the runtime trade-off opportunities in both algorithms and hardware to meet dynamically changing application performance targets and hardware constraints in real time.

Management Model Compression

GhostShiftAddNet: More Features from Energy-Efficient Operations

3 code implementations20 Sep 2021 Jia Bi, Jonathon Hare, Geoff V. Merrett

When compared to GhostNet, inference latency on the Jetson Nano is improved by 1. 3x and 2x on the GPU and CPU respectively.

Image Classification

Dynamic Transformer for Efficient Machine Translation on Embedded Devices

no code implementations17 Jul 2021 Hishan Parry, Lei Xun, Amin Sabet, Jia Bi, Jonathon Hare, Geoff V. Merrett

The new reduced design space results in a BLEU score increase of approximately 1% for sub-optimal models from the original design space, with a wide range for performance scaling between 0. 356s - 1. 526s for the GPU and 2. 9s - 7. 31s for the CPU.

Machine Translation Translation

Temporal Early Exits for Efficient Video Object Detection

no code implementations21 Jun 2021 Amin Sabet, Jonathon Hare, Bashir Al-Hashimi, Geoff V. Merrett

In this paper, we propose temporal early exits to reduce the computational complexity of per-frame video object detection.

Object object-detection +2

Optimising Resource Management for Embedded Machine Learning

1 code implementation8 May 2021 Lei Xun, Long Tran-Thanh, Bashir M Al-Hashimi, Geoff V. Merrett

Machine learning inference is increasingly being executed locally on mobile and embedded platforms, due to the clear advantages in latency, privacy and connectivity.

BIG-bench Machine Learning Management

Incremental Training and Group Convolution Pruning for Runtime DNN Performance Scaling on Heterogeneous Embedded Platforms

no code implementations8 May 2021 Lei Xun, Long Tran-Thanh, Bashir M Al-Hashimi, Geoff V. Merrett

Compared to the existing works, our approach can provide up to 2. 36x (energy) and 2. 73x (time) wider dynamic range with a 2. 4x smaller memory footprint at the same compression rate.

Dynamic-OFA: Runtime DNN Architecture Switching for Performance Scaling on Heterogeneous Embedded Platforms

1 code implementation8 May 2021 Wei Lou, Lei Xun, Amin Sabet, Jia Bi, Jonathon Hare, Geoff V. Merrett

However, the training process of such dynamic DNNs can be costly, since platform-aware models of different deployment scenarios must be retrained to become dynamic.

Cannot find the paper you are looking for? You can Submit a new open access paper.