no code implementations • 2 Dec 2024 • Chenghao Li, Yuke Zhang, Dake Chen, Jingqi Xu, Peter A. Beerel
In this paper, we address this gap by examining and mitigating the impact of the model structure, specifically the skip connections in the diffusion model's U-Net model.
no code implementations • 11 Nov 2024 • Jacob Huckelberry, Yuke Zhang, Allison Sansone, James Mickens, Peter A. Beerel, Vijay Janapa Reddi
Tiny Machine Learning (TinyML) systems, which enable machine learning inference on highly resource-constrained devices, are transforming edge computing but encounter unique security challenges.
no code implementations • 25 Sep 2024 • Md Abdullah-Al Kaiser, Sreetama Sarkar, Peter A. Beerel, Akhilesh R. Jaiswal, Gourav Datta
Moreover, they may not be suitable for real-time applications due to the long latency of modern CV networks that are deployed in the back-end.
no code implementations • 16 Jul 2024 • Sreetama Sarkar, Gourav Datta, Souvik Kundu, Kai Zheng, Chirayata Bhattacharyya, Peter A. Beerel
Video tasks are compute-heavy and thus pose a challenge when deploying in real-time applications, particularly for tasks that require state-of-the-art Vision Transformers (ViTs).
no code implementations • 8 Feb 2024 • Sreetama Sarkar, Souvik Kundu, Peter A. Beerel
Our experimental evaluations show that RLNet can yield models with up to 11. 14x fewer ReLUs, with accuracy close to the all-ReLU models, on clean, naturally perturbed, and gradient-based perturbed images.
no code implementations • 12 Dec 2023 • Gourav Datta, Zeyu Liu, James Diffenderfer, Bhavya Kailkhura, Peter A. Beerel
However, advanced ANN-to-SNN conversion approaches demonstrate that for lossless conversion, the number of SNN time steps must equal the number of quantization steps in the ANN activation function.
no code implementations • 2 Dec 2023 • Souvik Kundu, Rui-Jie Zhu, Akhilesh Jaiswal, Peter A. Beerel
Neuromorphic computing and, in particular, spiking neural networks (SNNs) have become an attractive alternative to deep neural networks for a broad range of signal processing applications, processing static and/or temporal inputs from different sensory modalities, including audio and vision sensors.
no code implementations • 28 Nov 2023 • Gourav Datta, Zeyu Liu, Anni Li, Peter A. Beerel
Recently proposed SNN training algorithms have significantly reduced the number of time steps (down to 1) for improved latency and energy efficiency, however, they target only convolutional neural networks (CNN).
no code implementations • 18 Oct 2023 • Sreetama Sarkar, Xinan Ye, Gourav Datta, Peter A. Beerel
This paper presents a comprehensive Deep Learning (DL) based on-line detection and correction approach, suitable for a wide range of pixel corruption rates.
1 code implementation • 13 Sep 2023 • Chenghao Li, Dake Chen, Yuke Zhang, Peter A. Beerel
While diffusion models demonstrate a remarkable capability for generating high-quality images, their tendency to `replicate' training data raises privacy concerns.
no code implementations • 8 Jun 2023 • Dake Chen, Christine Goins, Maxwell Waugaman, Georgios D. Dimou, Peter A. Beerel
In this paper, we describe and analyze an island-based random dynamic voltage scaling (iRDVS) approach to thwart power side-channel attacks.
no code implementations • 26 Apr 2023 • Souvik Kundu, Yuke Zhang, Dake Chen, Peter A. Beerel
Large number of ReLU and MAC operations of Deep neural networks make them ill-suited for latency and compute-efficient private inference.
no code implementations • 6 Apr 2023 • Md Abdullah-Al Kaiser, Gourav Datta, Sreetama Sarkar, Souvik Kundu, Zihan Yin, Manas Garg, Ajey P. Jacob, Peter A. Beerel, Akhilesh R. Jaiswal
The massive amounts of data generated by camera sensors motivate data processing inside pixel arrays, i. e., at the extreme-edge.
no code implementations • 17 Feb 2023 • Shashank Nag, Gourav Datta, Souvik Kundu, Nitin Chandrachoodan, Peter A. Beerel
Vision Transformer models, such as ViT, Swin Transformer, and Transformer-in-Transformer, have recently gained significant traction in computer vision tasks due to their ability to capture the global relation between features which leads to superior performance.
no code implementations • 23 Jan 2023 • Souvik Kundu, Shunlin Lu, Yuke Zhang, Jacqueline Liu, Peter A. Beerel
For a similar ReLU budget SENet can yield models with ~2. 32% improved classification accuracy, evaluated on CIFAR-100.
no code implementations • 22 Jan 2023 • Md Abdullah-Al Kaiser, Gourav Datta, Zixu Wang, Ajey P. Jacob, Peter A. Beerel, Akhilesh R. Jaiswal
Edge devices equipped with computer vision must deal with vast amounts of sensory data with limited computing resources.
no code implementations • ICCV 2023 • Yuke Zhang, Dake Chen, Souvik Kundu, Chenghao Li, Peter A. Beerel
Then, given our observation that external attention (EA) presents lower PI latency than widely-adopted self-attention (SA) at the cost of accuracy, we present a selective attention search (SAS) method to integrate the strength of EA and SA.
no code implementations • 27 Dec 2022 • Souvik Kundu, Sairam Sundaresan, Sharath Nittur Sridhar, Shunlin Lu, Han Tang, Peter A. Beerel
Existing deep neural networks (DNNs) that achieve state-of-the-art (SOTA) performance on both clean and adversarially-perturbed images rely on either activation or weight conditioned convolution operations.
no code implementations • 21 Dec 2022 • Gourav Datta, Zeyu Liu, Md Abdullah-Al Kaiser, Souvik Kundu, Joe Mathai, Zihan Yin, Ajey P. Jacob, Akhilesh R. Jaiswal, Peter A. Beerel
Although the overhead for the first layer MACs with direct encoding is negligible for deep SNNs and the CV processing is efficient using SNNs, the data transfer between the image sensors and the downstream processing costs significant bandwidth and may dominate the total energy.
no code implementations • 20 Dec 2022 • Gourav Datta, Zeyu Liu, Peter A. Beerel
Spiking Neural networks (SNN) have emerged as an attractive spatio-temporal computing paradigm for a wide range of low-power vision tasks.
no code implementations • 23 Oct 2022 • Gourav Datta, Haoqin Deng, Robert Aviles, Peter A. Beerel
We obtain test accuracy of 94. 75% with only 2 time steps with direct encoding on the GSC dataset with 4. 1x lower energy than an iso-architecture standard LSTM.
no code implementations • 11 Oct 2022 • Gourav Datta, Zeyu Liu, Zihan Yin, Linyu Sun, Akhilesh R. Jaiswal, Peter A. Beerel
However, direct inference on the raw images degrades the test accuracy due to the difference in covariance of the raw images captured by the image sensors compared to the ISP-processed images used for training.
no code implementations • 28 May 2022 • Gourav Datta, Souvik Kundu, Zihan Yin, Joe Mathai, Zeyu Liu, Zixu Wang, Mulin Tian, Shunlin Lu, Ravi T. Lakkireddy, Andrew Schmidt, Wael Abd-Almageed, Ajey P. Jacob, Akhilesh R. Jaiswal, Peter A. Beerel
The designs also reduce the sensor and total energy (obtained from in-house circuit simulations at Globalfoundries 22nm technology node) per frame by 5. 7x and 1. 14x, respectively.
no code implementations • 28 Mar 2022 • Souvik Kundu, Sairam Sundaresan, Massoud Pedram, Peter A. Beerel
In this paper, we present a fast learnable once-for-all adversarial training (FLOAT) algorithm, which instead of the existing FiLM-based conditioning, presents a unique weight conditioned learning that requires no additional layer, thereby incurring no significant increase in parameter count, training time, or network latency compared to standard adversarial training.
no code implementations • 11 Mar 2022 • Gourav Datta, Zihan Yin, Ajey Jacob, Akhilesh R. Jaiswal, Peter A. Beerel
Hyperspectral cameras generate a large amount of data due to the presence of hundreds of spectral bands as opposed to only three channels (red, green, and blue) in traditional cameras.
no code implementations • 7 Mar 2022 • Gourav Datta, Souvik Kundu, Zihan Yin, Ravi Teja Lakkireddy, Joe Mathai, Ajey Jacob, Peter A. Beerel, Akhilesh R. Jaiswal
Visual data in such cameras are usually captured in the form of analog voltages by a sensor pixel array, and then converted to the digital domain for subsequent AI processing using analog-to-digital converters (ADC).
no code implementations • 24 Dec 2021 • Souvik Kundu, Shikai Wang, Qirui Sun, Peter A. Beerel, Massoud Pedram
Compared to the baseline FP-32 models, BMPQ can yield models that have 15. 4x fewer parameter bits with a negligible drop in accuracy.
no code implementations • 22 Dec 2021 • Gourav Datta, Peter A. Beerel
SOTA training strategies for SNNs involve conversion from a non-spiking deep neural network (DNN).
no code implementations • 28 Oct 2021 • Yang Hu, Connor Imes, Xuanang Zhao, Souvik Kundu, Peter A. Beerel, Stephen P. Crago, John Paul N. Walters
We propose EdgePipe, a distributed framework for edge systems that uses pipeline parallelism to both speed up inference and enable running larger (and more accurate) models that otherwise cannot fit on single edge devices.
1 code implementation • ICCV 2021 • Souvik Kundu, Massoud Pedram, Peter A. Beerel
Low-latency deep spiking neural networks (SNNs) have become a promising alternative to conventional artificial neural networks (ANNs) because of their potential for increased energy efficiency on event-driven neuromorphic hardware.
no code implementations • 26 Jul 2021 • Gourav Datta, Souvik Kundu, Peter A. Beerel
This paper presents a training framework for low-latency energy-efficient SNNs that uses a hybrid encoding scheme at the input layer in which the analog pixel values of an image are directly applied during the first timestep and a novel variant of spike temporal coding is used during subsequent timesteps.
no code implementations • 26 Jul 2021 • Gourav Datta, Souvik Kundu, Akhilesh R. Jaiswal, Peter A. Beerel
However, the accurate processing of the spectral and spatial correlation between the bands requires the use of energy-expensive 3-D Convolutional Neural Networks (CNNs).
Computational Efficiency Hyperspectral Image Classification +1
no code implementations • 16 Jul 2021 • Souvik Kundu, Gourav Datta, Massoud Pedram, Peter A. Beerel
To evaluate the merits of our approach, we performed experiments with variants of VGG and ResNet, on both CIFAR-10 and CIFAR-100, and VGG16 on Tiny-ImageNet. The SNN models generated through the proposed technique yield SOTA compression ratios of up to 33. 4x with no significant drops in accuracy compared to baseline unpruned counterparts.
1 code implementation • 3 Nov 2020 • Souvik Kundu, Mahdi Nazemi, Peter A. Beerel, Massoud Pedram
This paper presents a dynamic network rewiring (DNR) method to generate pruned deep neural network (DNN) models that are robust against adversarial attacks yet maintain high accuracy on clean images.
1 code implementation • 6 Apr 2020 • Jennifer Paykin, Brian Huffman, Daniel M. Zimmerman, Peter A. Beerel
In this work we identify a counterexample to Cortadella et al.'s proof illustrating how their protocol can in fact lead to a violation of flow equivalence.
Logic in Computer Science
2 code implementations • 27 Mar 2020 • Sourya Dey, Saikrishna C. Kanala, Keith M. Chugg, Peter A. Beerel
In particular, we show the superiority of a greedy strategy and justify our choice of Bayesian optimization as the primary search methodology over random / grid search.
1 code implementation • 29 Jan 2020 • Souvik Kundu, Mahdi Nazemi, Massoud Pedram, Keith M. Chugg, Peter A. Beerel
We also compared the performance of our proposed architectures with that of ShuffleNet andMobileNetV2.
1 code implementation • 22 Oct 2019 • Arnab Sanyal, Peter A. Beerel, Keith M. Chugg
The high computational complexity associated with training deep neural networks limits online and real-time training on edge devices.
no code implementations • 2 Oct 2019 • Souvik Kundu, Saurav Prakash, Haleh Akrami, Peter A. Beerel, Keith M. Chugg
To explore the potential of this approach, we have experimented with two widely accepted datasets, CIFAR-10 and Tiny ImageNet, in sparse variants of both the ResNet18 and VGG16 architectures.
2 code implementations • 4 Dec 2018 • Sourya Dey, Kuan-Wen Huang, Peter A. Beerel, Keith M. Chugg
Neural networks have proven to be extremely powerful tools for modern artificial intelligence applications, but computational and storage complexity remain limiting factors.
2 code implementations • 11 Jul 2018 • Sourya Dey, Keith M. Chugg, Peter A. Beerel
The algorithm and datasets are open-source.
1 code implementation • 31 May 2018 • Sourya Dey, Diandian Chen, Zongyang Li, Souvik Kundu, Kuan-Wen Huang, Keith M. Chugg, Peter A. Beerel
We demonstrate an FPGA implementation of a parallel and reconfigurable architecture for sparse neural networks, capable of on-chip training and inference.
no code implementations • 18 Nov 2017 • Sourya Dey, Peter A. Beerel, Keith M. Chugg
We propose a class of interleavers for a novel deep neural network (DNN) architecture that uses algorithmically pre-determined, structured sparsity to significantly lower memory and computational requirements, and speed up training.
no code implementations • ICLR 2018 • Sourya Dey, Kuan-Wen Huang, Peter A. Beerel, Keith M. Chugg
We propose a novel way of reducing the number of parameters in the storage-hungry fully connected layers of a neural network by using pre-defined sparsity, where the majority of connections are absent prior to starting training.
no code implementations • 3 Nov 2017 • Sourya Dey, Yinan Shao, Keith M. Chugg, Peter A. Beerel
We propose a reconfigurable hardware architecture for deep neural networks (DNNs) capable of online training and inference, which uses algorithmically pre-determined, structured sparsity to significantly lower memory and computational requirements.