Search Results for author: Kartikeya Bhardwaj

Found 17 papers, 7 papers with code

Zero-Shot Neural Architecture Search: Challenges, Solutions, and Opportunities

1 code implementation5 Jul 2023 Guihong Li, Duc Hoang, Kartikeya Bhardwaj, Ming Lin, Zhangyang Wang, Radu Marculescu

Recently, zero-shot (or training-free) Neural Architecture Search (NAS) approaches have been proposed to liberate NAS from the expensive training process.

Neural Architecture Search

TIPS: Topologically Important Path Sampling for Anytime Neural Networks

no code implementations13 May 2023 Guihong Li, Kartikeya Bhardwaj, Yuedong Yang, Radu Marculescu

Anytime neural networks (AnytimeNNs) are a promising solution to adaptively adjust the model complexity at runtime under various hardware resource constraints.

ZiCo: Zero-shot NAS via Inverse Coefficient of Variation on Gradients

1 code implementation26 Jan 2023 Guihong Li, Yuedong Yang, Kartikeya Bhardwaj, Radu Marculescu

Based on this theoretical analysis, we propose a new zero-shot proxy, ZiCo, the first proxy that works consistently better than #Params.

Image Classification Neural Architecture Search

Restructurable Activation Networks

1 code implementation17 Aug 2022 Kartikeya Bhardwaj, James Ward, Caleb Tung, Dibakar Gope, Lingchuan Meng, Igor Fedorov, Alex Chalfin, Paul Whatmough, Danny Loh

To address this question, we propose a new paradigm called Restructurable Activation Networks (RANs) that manipulate the amount of non-linearity in models to improve their hardware-awareness and efficiency.

object-detection Object Detection

Collapsible Linear Blocks for Super-Efficient Super Resolution

3 code implementations17 Mar 2021 Kartikeya Bhardwaj, Milos Milosavljevic, Liam O'Neil, Dibakar Gope, Ramon Matas, Alex Chalfin, Naveen Suda, Lingchuan Meng, Danny Loh

Our results highlight the challenges faced by super resolution on AI accelerators and demonstrate that SESR is significantly faster (e. g., 6x-8x higher FPS) than existing models on mobile-NPU.

4k 8k +1

On the relationship between topology and gradient propagation in deep networks

no code implementations1 Jan 2021 Kartikeya Bhardwaj, Guihong Li, Radu Marculescu

(ii) Can certain topological characteristics of deep networks indicate a priori (i. e., without training) which models, with a different number of parameters/FLOPS/layers, achieve a similar accuracy?

New Directions in Distributed Deep Learning: Bringing the Network at Forefront of IoT Design

no code implementations25 Aug 2020 Kartikeya Bhardwaj, Wei Chen, Radu Marculescu

In this paper, we first highlight three major challenges to large-scale adoption of deep learning at the edge: (i) Hardware-constrained IoT devices, (ii) Data security and privacy in the IoT era, and (iii) Lack of network-aware deep learning algorithms for distributed inference across multiple IoT devices.

Federated Learning

FedMAX: Mitigating Activation Divergence for Accurate and Communication-Efficient Federated Learning

1 code implementation7 Apr 2020 Wei Chen, Kartikeya Bhardwaj, Radu Marculescu

In this paper, we identify a new phenomenon called activation-divergence which occurs in Federated Learning (FL) due to data heterogeneity (i. e., data being non-IID) across multiple users.

Federated Learning

EdgeAI: A Vision for Deep Learning in IoT Era

no code implementations23 Oct 2019 Kartikeya Bhardwaj, Naveen Suda, Radu Marculescu

The significant computational requirements of deep learning present a major bottleneck for its large-scale adoption on hardware-constrained IoT-devices.

How does topology influence gradient propagation and model performance of deep networks with DenseNet-type skip connections?

2 code implementations CVPR 2021 Kartikeya Bhardwaj, Guihong Li, Radu Marculescu

In this paper, we reveal that the topology of the concatenation-type skip connections is closely related to the gradient propagation which, in turn, enables a predictable behavior of DNNs' test performance.

Model Compression

Memory- and Communication-Aware Model Compression for Distributed Deep Learning Inference on IoT

no code implementations26 Jul 2019 Kartikeya Bhardwaj, Chingyi Lin, Anderson Sartor, Radu Marculescu

Therefore, we propose Network of Neural Networks (NoNN), a new distributed IoT learning paradigm that compresses a large pretrained 'teacher' deep network into several disjoint and highly-compressed 'student' modules, without loss of accuracy.

Image Classification Model Compression

A Dynamic Network and Representation LearningApproach for Quantifying Economic Growth fromSatellite Imagery

no code implementations1 Dec 2018 Jiqian Dong, Gopaljee Atulya, Kartikeya Bhardwaj, Radu Marculescu

To this end, we propose a new network science- and representation learning-based approach that can quantify economic indicators and visualize the growth of various regions.

Representation Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.