no code implementations • 12 Mar 2024 • Cristian Cioflan, Lukas Cavigelli, Manuele Rusci, Miguel de Prado, Luca Benini
Keyword spotting accuracy degrades when neural networks are exposed to noisy environments.
no code implementations • 12 Mar 2024 • Cristian Cioflan, Lukas Cavigelli, Luca Benini
Keyword spotting systems for always-on TinyML-constrained applications require on-site tuning to boost the accuracy of offline trained classifiers when deployed in unseen inference conditions.
1 code implementation • 16 Nov 2023 • Jannis Schönleber, Lukas Cavigelli, Renzo Andri, Matteo Perotti, Luca Benini
From classical HPC to deep learning, MatMul is at the heart of today's computing.
1 code implementation • NeurIPS 2023 • Antoine Scardigli, Lukas Cavigelli, Lorenz K. Müller
Monte-Carlo path tracing is a powerful technique for realistic image synthesis but suffers from high levels of noise at low sample counts, limiting its use in real-time applications.
no code implementations • 26 Sep 2022 • Renzo Andri, Beatrice Bussolino, Antonio Cipolletta, Lukas Cavigelli, Zhe Wang
The Winograd-enhanced DSA achieves up to 1. 85x gain in energy efficiency and up to 1. 83x end-to-end speed-up for state-of-the-art segmentation and detection networks.
1 code implementation • IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS) 2022 • Cristian Cioflan, Lukas Cavigelli, Manuele Rusci, Miguel de Prado, Luca Benini
The accuracy of a keyword spotting model deployed on embedded devices often degrades when the system is exposed to real environments with significant noise.
no code implementations • 14 Feb 2022 • Gianna Paulin, Francesco Conti, Lukas Cavigelli, Luca Benini
For quantifying the overall system power, including I/O power, we built Vau da Muntanialas, to the best of our knowledge, the first demonstration of a systolic multi-chip-on-PCB array of RNN accelerator.
no code implementations • 18 Dec 2021 • Xiaying Wang, Lukas Cavigelli, Tibor Schneider, Luca Benini
Motor imagery brain--machine interfaces enable us to control machines by merely thinking of performing a motor action.
no code implementations • 4 May 2021 • Xavier Timoneda, Lukas Cavigelli
Logic optimization is an NP-hard problem commonly approached through hand-engineered heuristics.
1 code implementation • 25 Mar 2021 • Thorir Mar Ingolfsson, Xiaying Wang, Michael Hersche, Alessio Burrello, Lukas Cavigelli, Luca Benini
With 9. 91 GMAC/s/W, it is 23. 0 times more energy-efficient and 46. 85 times faster than an implementation on the ARM Cortex M4F (0. 43 GMAC/s/W).
1 code implementation • 22 Feb 2021 • Xiaying Wang, Tibor Schneider, Michael Hersche, Lukas Cavigelli, Luca Benini
With Motor-Imagery (MI) Brain--Machine Interfaces (BMIs) we may control machines by merely thinking of performing a motor action.
no code implementations • 12 Jan 2021 • Gianmarco Cerutti, Renzo Andri, Lukas Cavigelli, Michele Magno, Elisabetta Farella, Luca Benini
This BNN reaches a 77. 9% accuracy, just 7% lower than the full-precision version, with 58 kB (7. 2 times less) for the weights and 262 kB (2. 4 times less) memory in total.
no code implementations • 3 Nov 2020 • Moritz Scherer, Georg Rutishauser, Lukas Cavigelli, Luca Benini
We present a 3. 1 POp/s/W fully digital hardware accelerator for ternary neural networks.
Hardware Architecture
1 code implementation • 31 May 2020 • Thorir Mar Ingolfsson, Michael Hersche, Xiaying Wang, Nobuaki Kobayashi, Lukas Cavigelli, Luca Benini
Experimental results on the BCI Competition IV-2a dataset show that EEG-TCNet achieves 77. 35% classification accuracy in 4-class MI.
no code implementations • 12 May 2020 • Renzo Andri, Geethan Karunaratne, Lukas Cavigelli, Luca Benini
Furthermore, it can perform inference on a binarized ResNet-18 trained with 8-bases Group-Net to achieve a 67. 5% Top-1 accuracy with only 3. 0 mJ/frame -- at an accuracy drop of merely 1. 8% from the full-precision ResNet-18.
1 code implementation • 24 Apr 2020 • Tibor Schneider, Xiaying Wang, Michael Hersche, Lukas Cavigelli, Luca Benini
We quantize weights and activations to 8-bit fixed-point with a negligible accuracy loss of 0. 4% on 4-class MI, and present an energy-efficient hardware-aware implementation on the Mr. Wolf parallel ultra-low power (PULP) System-on-Chip (SoC) by utilizing its custom RISC-V ISA extensions and 8-core compute cluster.
no code implementations • 28 Feb 2020 • Michele Magno, Xiaying Wang, Manuel Eggimann, Lukas Cavigelli, Luca Benini
This work presents InfiniWolf, a novel multi-sensor smartwatch that can achieve self-sustainability exploiting thermal and solar energy harvesting, performing computationally high demanding tasks.
no code implementations • 4 Jan 2020 • Lukas Cavigelli, Luca Benini
We present Random Partition Relaxation (RPR), a method for strong quantization of neural networks weight to binary (+1/-1) and ternary (+1/0/-1) values.
no code implementations • 10 Dec 2019 • Xiaying Wang, Lukas Cavigelli, Manuel Eggimann, Michele Magno, Luca Benini
Synthetic aperture radar (SAR) data is becoming increasingly available to a wide range of users through commercial service providers with resolutions reaching 0. 5m/px.
1 code implementation • 8 Nov 2019 • Xiaying Wang, Michele Magno, Lukas Cavigelli, Luca Benini
The growing number of low-power smart devices in the Internet of Things is coupled with the concept of "Edge Computing", that is moving some of the intelligence, especially machine learning, towards the edge of the network.
2 code implementations • 30 Aug 2019 • Lukas Cavigelli, Georg Rutishauser, Luca Benini
In the wake of the success of convolutional neural networks in image classification, object recognition, speech recognition, etc., the demand for deploying these compute-intensive ML models on embedded and mobile systems with tight power and energy constraints at low cost, as well as for boosting throughput in data centers, is growing rapidly.
1 code implementation • 24 May 2019 • Matteo Spallanzani, Lukas Cavigelli, Gian Paolo Leonardi, Marko Bertogna, Luca Benini
We present a theoretical and experimental investigation of the quantization problem for artificial neural networks.
1 code implementation • 1 Oct 2018 • Lukas Cavigelli, Luca Benini
After the tremendous success of convolutional neural networks in image classification, object detection, speech recognition, etc., there is now rising demand for deployment of these compute-intensive ML models on tightly power constrained embedded and mobile systems at low cost as well as for pushing the throughput in data centers.
2 code implementations • 15 Aug 2018 • Lukas Cavigelli, Luca Benini
The last few years have brought advances in computer vision at an amazing pace, grounded on new findings in deep neural network construction and training as well as the availability of large labeled datasets.
2 code implementations • 18 Jun 2018 • Michael Hersche, Tino Rellstab, Pasquale Davide Schiavone, Lukas Cavigelli, Luca Benini, Abbas Rahimi
Accurate, fast, and reliable multiclass classification of electroencephalography (EEG) signals is a challenging task towards the development of motor imagery brain-computer interface (MI-BCI) systems.
no code implementations • 5 Mar 2018 • Renzo Andri, Lukas Cavigelli, Davide Rossi, Luca Benini
Deep neural networks have achieved impressive results in computer vision and machine learning.
no code implementations • 5 Mar 2018 • Andrawes Al Bahou, Geethan Karunaratne, Renzo Andri, Lukas Cavigelli, Luca Benini
Deploying state-of-the-art CNNs requires power-hungry processors and off-chip memory.
no code implementations • 21 Nov 2017 • Manuele Rusci, Lukas Cavigelli, Luca Benini
Design automation in general, and in particular logic synthesis, can play a key role in enabling the design of application-specific Binarized Neural Networks (BNN).
no code implementations • 15 Nov 2017 • Francesco Conti, Lukas Cavigelli, Gianna Paulin, Igor Susmelj, Luca Benini
Recurrent neural networks (RNNs) are state-of-the-art in voice awareness/understanding and speech recognition.
no code implementations • 28 Sep 2017 • Matthias Meyer, Lukas Cavigelli, Lothar Thiele
Wireless distributed systems as used in sensor networks, Internet-of-Things and cyber-physical systems, impose high requirements on resource efficiency.
1 code implementation • 14 Apr 2017 • Lukas Cavigelli, Philippe Degen, Luca Benini
Extracting per-frame features using convolutional neural networks for real-time processing of video data is currently mainly performed on powerful GPU-accelerated workstations and compute clusters.
no code implementations • NeurIPS 2017 • Eirikur Agustsson, Fabian Mentzer, Michael Tschannen, Lukas Cavigelli, Radu Timofte, Luca Benini, Luc van Gool
We present a new approach to learn compressible representations in deep architectures with an end-to-end training strategy.
1 code implementation • 22 Nov 2016 • Lukas Cavigelli, Pascal Hager, Luca Benini
Lossy image compression algorithms are pervasively used to reduce the size of images transmitted over the web and recorded on data storage media.
no code implementations • 9 Nov 2016 • Lukas Cavigelli, Dominic Bernath, Michele Magno, Luca Benini
The required communication links and archiving of the video data are still expensive and this setup excludes preemptive actions to respond to imminent threats.
no code implementations • 26 Sep 2016 • Michael Tschannen, Lukas Cavigelli, Fabian Mentzer, Thomas Wiatowski, Luca Benini
We propose a highly structured neural network architecture for semantic segmentation with an extremely small model size, suitable for low-power embedded and mobile platforms.
no code implementations • 17 Jun 2016 • Renzo Andri, Lukas Cavigelli, Davide Rossi, Luca Benini
Convolutional neural networks (CNNs) have revolutionized the world of computer vision over the last few years, pushing image classification beyond human accuracy.
no code implementations • 14 Dec 2015 • Lukas Cavigelli, Luca Benini
An ever increasing number of computer vision and image/video processing challenges are being approached using deep convolutional neural networks, obtaining state-of-the-art results in object recognition and detection, semantic segmentation, action recognition, optical flow and superresolution.