1 code implementation • 17 Oct 2023 • Hongxiang Fan, Stylianos I. Venieris, Alexandros Kouris, Nicholas D. Lane
Running multiple deep neural networks (DNNs) in parallel has become an emerging workload in both edge devices, such as mobile phones where multiple tasks serve a single user for daily activities, and data centers, where various requests are raised from millions of users, as seen with large language models.
no code implementations • 19 Oct 2022 • Stefanos Laskaridis, Stylianos I. Venieris, Alexandros Kouris, Rui Li, Nicholas D. Lane
In the last decade, Deep Learning has rapidly infiltrated the consumer end, mainly thanks to hardware acceleration across devices.
no code implementations • 27 Sep 2022 • Alexandros Kouris, Stylianos I. Venieris, Stefanos Laskaridis, Nicholas D. Lane
With deep neural networks (DNNs) emerging as the backbone in a multitude of computer vision tasks, their adoption in real-world applications broadens continuously.
no code implementations • 20 Sep 2022 • Hongxiang Fan, Thomas Chau, Stylianos I. Venieris, Royson Lee, Alexandros Kouris, Wayne Luk, Nicholas D. Lane, Mohamed S. Abdelfattah
By jointly optimizing the algorithm and hardware, our FPGA-based butterfly accelerator achieves 14. 2 to 23. 2 times speedup over state-of-the-art accelerators normalized to the same computational budget.
no code implementations • 9 Jun 2021 • Stefanos Laskaridis, Alexandros Kouris, Nicholas D. Lane
DNNs are becoming less and less over-parametrised due to recent advances in efficient model design, through careful hand-crafted or NAS-based methods.
no code implementations • 7 Jun 2021 • Alexandros Kouris, Stylianos I. Venieris, Stefanos Laskaridis, Nicholas D. Lane
At the same time, the heterogeneous capabilities of the target platforms and the diverse constraints of different applications require the design and training of multiple target-specific segmentation models, leading to excessive maintenance costs.
no code implementations • 2 May 2019 • Alexandros Kouris, Stylianos I. Venieris, Michail Rizakis, Christos-Savvas Bouganis
The need to recognise long-term dependencies in sequential data such as video streams has made Long Short-Term Memory (LSTM) networks a prominent Artificial Intelligence model for many emerging applications.
no code implementations • 13 Jul 2018 • Alexandros Kouris, Stylianos I. Venieris, Christos-Savvas Bouganis
This work presents CascadeCNN, an automated toolflow that pushes the quantisation limits of any given CNN model, aiming to perform high-throughput inference.
no code implementations • 22 Jun 2018 • Stylianos I. Venieris, Alexandros Kouris, Christos-Savvas Bouganis
Recently, Deep Neural Networks (DNNs) have emerged as the dominant model across various AI applications.
no code implementations • 22 May 2018 • Alexandros Kouris, Stylianos I. Venieris, Christos-Savvas Bouganis
This work presents CascadeCNN, an automated toolflow that pushes the quantisation limits of any given CNN model, to perform high-throughput inference by exploiting the computation time-accuracy trade-off.
no code implementations • 15 Mar 2018 • Stylianos I. Venieris, Alexandros Kouris, Christos-Savvas Bouganis
In the past decade, Convolutional Neural Networks (CNNs) have demonstrated state-of-the-art performance in various Artificial Intelligence tasks.
no code implementations • 7 Jan 2018 • Michalis Rizakis, Stylianos I. Venieris, Alexandros Kouris, Christos-Savvas Bouganis
Recurrent Neural Networks and in particular Long Short-Term Memory (LSTM) networks have demonstrated state-of-the-art accuracy in several emerging Artificial Intelligence tasks.