Search Results for author: Viktor Prasanna

Found 34 papers, 14 papers with code

GCV-Turbo: End-to-end Acceleration of GNN-based Computer Vision Tasks on FPGA

no code implementations10 Apr 2024 Bingyi Zhang, Rajgopal Kannan, Carl Busart, Viktor Prasanna

Moreover, GCV-Turbo supports the execution of the standalone CNNs or GNNs, achieving performance comparable to that of state-of-the-art CNN (GNN) accelerators for widely used CNN-only (GNN-only) models.

TabConv: Low-Computation CNN Inference via Table Lookups

1 code implementation8 Apr 2024 Neelesh Gupta, Narayanan Kannan, Pengmiao Zhang, Viktor Prasanna

TabConv preserves over 93% of the original model's performance while reducing arithmetic operations by 36. 5%, 25. 8%, and 99. 4% for ResNet-18 on CIFAR-10, CIFAR-100, and MNIST, respectively, 35. 6% and 99. 3% for ResNet-34 on CIFAR-10 and MNIST, and 98. 9% for NIN on MNIST, achieving low-computation inference.

VTR: An Optimized Vision Transformer for SAR ATR Acceleration on FPGA

no code implementations6 Apr 2024 Sachini Wickramasinghe, Dhruv Parikh, Bingyi Zhang, Rajgopal Kannan, Viktor Prasanna, Carl Busart

We directly train this model on SAR datasets which have limited training samples to evaluate its effectiveness for SAR ATR applications.

FACTUAL: A Novel Framework for Contrastive Learning Based Robust SAR Image Classification

no code implementations4 Apr 2024 Xu Wang, Tian Ye, Rajgopal Kannan, Viktor Prasanna

FACTUAL consists of two components: (1) Differing from existing works, a novel perturbation scheme that incorporates realistic physical adversarial attacks (such as OTSA) to build a supervised adversarial pre-training network.

Contrastive Learning Image Classification

Uncertainty-Aware SAR ATR: Defending Against Adversarial Attacks via Bayesian Neural Networks

no code implementations27 Mar 2024 Tian Ye, Rajgopal Kannan, Viktor Prasanna, Carl Busart

Adversarial attacks have demonstrated the vulnerability of Machine Learning (ML) image classifiers in Synthetic Aperture Radar (SAR) Automatic Target Recognition (ATR) systems.

Adversarial Attack Decision Making +1

Accelerating ViT Inference on FPGA through Static and Dynamic Pruning

no code implementations21 Mar 2024 Dhruv Parikh, Shouyi Li, Bingyi Zhang, Rajgopal Kannan, Carl Busart, Viktor Prasanna

For algorithm design, we systematically combine a hardware-aware structured block-pruning method for pruning model parameters and a dynamic token pruning method for removing unimportant token vectors.

PaCKD: Pattern-Clustered Knowledge Distillation for Compressing Memory Access Prediction Models

1 code implementation21 Feb 2024 Neelesh Gupta, Pengmiao Zhang, Rajgopal Kannan, Viktor Prasanna

Deep neural networks (DNNs) have proven to be effective models for accurate Memory Access Prediction (MAP), a critical task in mitigating memory latency through data prefetching.

Image Classification Knowledge Distillation

ME-ViT: A Single-Load Memory-Efficient FPGA Accelerator for Vision Transformers

no code implementations15 Feb 2024 Kyle Marino, Pengmiao Zhang, Viktor Prasanna

We evaluate ME-ViT on systolic array sizes of 32 and 16, achieving up to a 9. 22$\times$ and 17. 89$\times$ overall improvement in memory bandwidth, and a 2. 16$\times$ improvement in throughput per DSP for both designs over state-of-the-art ViT accelerators on FPGA.

TASER: Temporal Adaptive Sampling for Fast and Accurate Dynamic Graph Representation Learning

1 code implementation8 Feb 2024 Gangda Deng, Hongkuan Zhou, Hanqing Zeng, Yinglong Xia, Christopher Leung, Jianbo Li, Rajgopal Kannan, Viktor Prasanna

Recently, Temporal Graph Neural Networks (TGNNs) have demonstrated state-of-the-art performance in various high-impact applications, including fraud detection and content recommendation.

Denoising Fraud Detection +1

A Single Graph Convolution Is All You Need: Efficient Grayscale Image Classification

1 code implementation1 Feb 2024 Jacob Fein-Ashley, Tian Ye, Sachini Wickramasinghe, Bingyi Zhang, Rajgopal Kannan, Viktor Prasanna

Our experimental results on benchmark grayscale image datasets demonstrate the effectiveness of the proposed model, achieving vastly lower latency (up to 16$\times$ less) and competitive or leading performance compared to other state-of-the-art image classification models on various domain-specific grayscale image classification datasets.

Image Classification Medical Image Classification

PAHD: Perception-Action based Human Decision Making using Explainable Graph Neural Networks on SAR Images

no code implementations5 Jan 2024 Sasindu Wijeratne, Bingyi Zhang, Rajgopal Kannan, Viktor Prasanna, Carl Busart

This detailed information includes the SAR image features that contributed to the classification, the classification confidence, and the probability of the identified object being classified as a different object type or class.

Decision Making Object

Benchmarking Deep Learning Classifiers for SAR Automatic Target Recognition

no code implementations12 Dec 2023 Jacob Fein-Ashley, Tian Ye, Rajgopal Kannan, Viktor Prasanna, Carl Busart

Synthetic Aperture Radar SAR Automatic Target Recognition ATR is a key technique of remote-sensing image recognition which can be supported by deep neural networks The existing works of SAR ATR mostly focus on improving the accuracy of the target recognition while ignoring the systems performance in terms of speed and storage which is critical to real-world applications of SAR ATR For decision-makers aiming to identify a proper deep learning model to deploy in a SAR ATR system it is important to understand the performance of different candidate deep learning models and determine the best model accordingly This paper comprehensively benchmarks several advanced deep learning models for SAR ATR with multiple distinct SAR imagery datasets Specifically we train and test five SAR image classifiers based on Residual Neural Networks ResNet18 ResNet34 ResNet50 Graph Neural Network GNN and Vision Transformer for Small-Sized Datasets (SS-ViT) We select three datasets MSTAR GBSAR and SynthWakeSAR that offer heterogeneity We evaluate and compare the five classifiers concerning their classification accuracy runtime performance in terms of inference throughput and analytical performance in terms of number of parameters number of layers model size and number of operations Experimental results show that the GNN classifier outperforms with respect to throughput and latency However it is also shown that no clear model winner emerges from all of our chosen metrics and a one model rules all case is doubtful in the domain of SAR ATR

Benchmarking

Realistic Scatterer Based Adversarial Attacks on SAR Image Classifiers

no code implementations5 Dec 2023 Tian Ye, Rajgopal Kannan, Viktor Prasanna, Carl Busart, Lance Kaplan

Instead, adversarial attacks should be able to be implemented by physical actions, for example, placing additional false objects as scatterers around the on-ground target to perturb the SAR image and fool the SAR ATR.

Adversarial Attack

Characterizing Speed Performance of Multi-Agent Reinforcement Learning

no code implementations13 Sep 2023 Samuel Wiggins, Yuan Meng, Rajgopal Kannan, Viktor Prasanna

Multi-Agent Reinforcement Learning (MARL) has achieved significant success in large-scale AI systems and big-data applications such as smart grids, surveillance, etc.

Multi-agent Reinforcement Learning reinforcement-learning

Exploiting On-chip Heterogeneity of Versal Architecture for GNN Inference Acceleration

no code implementations4 Aug 2023 Paul Chen, Pavan Manjunath, Sasindu Wijeratne, Bingyi Zhang, Viktor Prasanna

To exploit data sparsity during inference, we devise a runtime kernel mapping strategy that dynamically assigns computation tasks to the PL and AIE based on data sparsity.

DistTGL: Distributed Memory-Based Temporal Graph Neural Network Training

no code implementations14 Jul 2023 Hongkuan Zhou, Da Zheng, Xiang Song, George Karypis, Viktor Prasanna

Evenworse, the tremendous overhead to synchronize the node memory make it impractical to be deployed to distributed GPU clusters.

Graph Representation Learning

Graph Neural Network for Accurate and Low-complexity SAR ATR

no code implementations11 May 2023 Bingyi Zhang, Sasindu Wijeratne, Rajgopal Kannan, Viktor Prasanna, Carl Busart

In this work, we propose a graph neural network (GNN) model to achieve accurate and low-latency SAR ATR.

Accurate, Low-latency, Efficient SAR Automatic Target Recognition on FPGA

no code implementations4 Jan 2023 Bingyi Zhang, Rajgopal Kannan, Viktor Prasanna, Carl Busart

Compared with the state-of-the-art CNNs, the proposed GNN achieves comparable accuracy with $1/3258$ computation cost and $1/83$ model size.

Towards Programmable Memory Controller for Tensor Decomposition

no code implementations17 Jul 2022 Sasindu Wijeratne, Ta-Yang Wang, Rajgopal Kannan, Viktor Prasanna

Implementing accelerators on Field Programmable Gate Array (FPGA) for kernels such as MTTKRP is attractive due to the energy efficiency and the inherent parallelism of FPGA.

Tensor Decomposition

Model-Architecture Co-Design for High Performance Temporal GNN Inference on FPGA

1 code implementation10 Mar 2022 Hongkuan Zhou, Bingyi Zhang, Rajgopal Kannan, Viktor Prasanna, Carl Busart

Taking advantage of the model optimizations, we propose a principled hardware architecture using batching, pipelining, and prefetching techniques to further improve the performance.

Knowledge Distillation

Decoupling the Depth and Scope of Graph Neural Networks

1 code implementation NeurIPS 2021 Hanqing Zeng, Muhan Zhang, Yinglong Xia, Ajitesh Srivastava, Andrey Malevich, Rajgopal Kannan, Viktor Prasanna, Long Jin, Ren Chen

We propose a design principle to decouple the depth and scope of GNNs -- to generate representation of a target entity (i. e., a node or an edge), we first extract a localized subgraph as the bounded-size scope, and then apply a GNN of arbitrary depth on top of the subgraph.

Link Prediction Node Classification +1

Reconfigurable Low-latency Memory System for Sparse Matricized Tensor Times Khatri-Rao Product on FPGA

no code implementations18 Sep 2021 Sasindu Wijeratne, Rajgopal Kannan, Viktor Prasanna

This paper focuses on a multi-faceted memory system, which explores the spatial and temporal locality of the data structures of MTTKRP.

Tensor Decomposition

SeDyT: A General Framework for Multi-Step Event Forecasting via Sequence Modeling on Dynamic Entity Embeddings

1 code implementation9 Sep 2021 Hongkuan Zhou, James Orme-Rogers, Rajgopal Kannan, Viktor Prasanna

SeDyT consists of two components: a Temporal Graph Neural Network that generates dynamic entity embeddings in the past and a sequence model that predicts the entity embeddings in the future.

Entity Embeddings Knowledge Graphs

Programmable FPGA-based Memory Controller

no code implementations21 Aug 2021 Sasindu Wijeratne, Sanket Pattnaik, Zhiyu Chen, Rajgopal Kannan, Viktor Prasanna

Since developing memory controllers for different applications is time-consuming, this paper introduces a modular and programmable memory controller that can be configured for different target applications on available hardware resources.

Scheduling

Accelerating Large Scale Real-Time GNN Inference using Channel Pruning

1 code implementation10 May 2021 Hongkuan Zhou, Ajitesh Srivastava, Hanqing Zeng, Rajgopal Kannan, Viktor Prasanna

In this paper, we propose to accelerate GNN inference by pruning the dimensions in each layer with negligible accuracy loss.

Node Classification Spam detection

BRAC+: Going Deeper with Behavior Regularized Offline Reinforcement Learning

no code implementations1 Jan 2021 Chi Zhang, Sanmukh Rao Kuppannagari, Viktor Prasanna

The goal of Offline Reinforcement Learning (RL) is to address this problem by learning effective policies using previously collected datasets.

Offline RL reinforcement-learning +1

Deep Graph Neural Networks with Shallow Subgraph Samplers

2 code implementations2 Dec 2020 Hanqing Zeng, Muhan Zhang, Yinglong Xia, Ajitesh Srivastava, Andrey Malevich, Rajgopal Kannan, Viktor Prasanna, Long Jin, Ren Chen

We propose a simple "deep GNN, shallow sampler" design principle to improve both the GNN accuracy and efficiency -- to generate representation of a target node, we use a deep GNN to pass messages only within a shallow, localized subgraph.

Accurate, Efficient and Scalable Training of Graph Neural Networks

2 code implementations5 Oct 2020 Hanqing Zeng, Hongkuan Zhou, Ajitesh Srivastava, Rajgopal Kannan, Viktor Prasanna

For feature propagation within subgraphs, we improve cache utilization and reduce DRAM traffic by data partitioning.

Graph Sampling

GraphACT: Accelerating GCN Training on CPU-FPGA Heterogeneous Platforms

1 code implementation31 Dec 2019 Hanqing Zeng, Viktor Prasanna

We first analyze the computation and communication characteristics of various GCN training algorithms, and select a subgraph-based algorithm that is well suited for hardware execution.

Representation Learning

SPEC2: SPECtral SParsE CNN Accelerator on FPGAs

no code implementations16 Oct 2019 Yue Niu, Hanqing Zeng, Ajitesh Srivastava, Kartik Lakhotia, Rajgopal Kannan, Yanzhi Wang, Viktor Prasanna

On the other hand, weight pruning techniques address the redundancy in model parameters by converting dense convolutional kernels into sparse ones.

Accurate, Efficient and Scalable Graph Embedding

2 code implementations28 Oct 2018 Hanqing Zeng, Hongkuan Zhou, Ajitesh Srivastava, Rajgopal Kannan, Viktor Prasanna

However, a major challenge is to reduce the complexity of layered GCNs and make them parallelizable and scalable on very large graphs -- state-of the art techniques are unable to achieve scalability without losing accuracy and efficiency.

Clustering Graph Embedding +2

Accelerating PageRank using Partition-Centric Processing

1 code implementation21 Sep 2017 Kartik Lakhotia, Rajgopal Kannan, Viktor Prasanna

The traditional PageRank implementation generates fine granularity random memory accesses resulting in large amount of wasteful DRAM traffic and poor bandwidth utilization.

Distributed, Parallel, and Cluster Computing Data Structures and Algorithms Performance

Cannot find the paper you are looking for? You can Submit a new open access paper.