no code implementations • 26 Mar 2024 • Kartikeya Bhardwaj, Nilesh Prasad Pandey, Sweta Priyadarshi, Kyunggeun Lee, Jun Ma, Harris Teague
Large generative models such as large language models (LLMs) and diffusion models have revolutionized the fields of NLP and computer vision respectively.
no code implementations • 26 Sep 2023 • Kartikeya Bhardwaj, Hsin-Pai Cheng, Sweta Priyadarshi, Zhuojin Li
To solve the problem, we propose a novel bias correction on ZiCo, called ZiCo-BC.
1 code implementation • 5 Jul 2023 • Guihong Li, Duc Hoang, Kartikeya Bhardwaj, Ming Lin, Zhangyang Wang, Radu Marculescu
Recently, zero-shot (or training-free) Neural Architecture Search (NAS) approaches have been proposed to liberate NAS from the expensive training process.
no code implementations • 13 May 2023 • Guihong Li, Kartikeya Bhardwaj, Yuedong Yang, Radu Marculescu
Anytime neural networks (AnytimeNNs) are a promising solution to adaptively adjust the model complexity at runtime under various hardware resource constraints.
1 code implementation • 26 Jan 2023 • Guihong Li, Yuedong Yang, Kartikeya Bhardwaj, Radu Marculescu
Based on this theoretical analysis, we propose a new zero-shot proxy, ZiCo, the first proxy that works consistently better than #Params.
1 code implementation • 17 Aug 2022 • Kartikeya Bhardwaj, James Ward, Caleb Tung, Dibakar Gope, Lingchuan Meng, Igor Fedorov, Alex Chalfin, Paul Whatmough, Danny Loh
To address this question, we propose a new paradigm called Restructurable Activation Networks (RANs) that manipulate the amount of non-linearity in models to improve their hardware-awareness and efficiency.
1 code implementation • 29 Dec 2021 • Kartikeya Bhardwaj, Dibakar Gope, James Ward, Paul Whatmough, Danny Loh
Autonomous systems are highly vulnerable to a variety of adversarial attacks on Deep Neural Networks (DNNs).
3 code implementations • 17 Mar 2021 • Kartikeya Bhardwaj, Milos Milosavljevic, Liam O'Neil, Dibakar Gope, Ramon Matas, Alex Chalfin, Naveen Suda, Lingchuan Meng, Danny Loh
Our results highlight the challenges faced by super resolution on AI accelerators and demonstrate that SESR is significantly faster (e. g., 6x-8x higher FPS) than existing models on mobile-NPU.
no code implementations • 1 Jan 2021 • Kartikeya Bhardwaj, Guihong Li, Radu Marculescu
(ii) Can certain topological characteristics of deep networks indicate a priori (i. e., without training) which models, with a different number of parameters/FLOPS/layers, achieve a similar accuracy?
no code implementations • 25 Aug 2020 • Kartikeya Bhardwaj, Wei Chen, Radu Marculescu
In this paper, we first highlight three major challenges to large-scale adoption of deep learning at the edge: (i) Hardware-constrained IoT devices, (ii) Data security and privacy in the IoT era, and (iii) Lack of network-aware deep learning algorithms for distributed inference across multiple IoT devices.
1 code implementation • 7 Apr 2020 • Wei Chen, Kartikeya Bhardwaj, Radu Marculescu
In this paper, we identify a new phenomenon called activation-divergence which occurs in Federated Learning (FL) due to data heterogeneity (i. e., data being non-IID) across multiple users.
no code implementations • 23 Oct 2019 • Kartikeya Bhardwaj, Naveen Suda, Radu Marculescu
The significant computational requirements of deep learning present a major bottleneck for its large-scale adoption on hardware-constrained IoT-devices.
2 code implementations • CVPR 2021 • Kartikeya Bhardwaj, Guihong Li, Radu Marculescu
In this paper, we reveal that the topology of the concatenation-type skip connections is closely related to the gradient propagation which, in turn, enables a predictable behavior of DNNs' test performance.
no code implementations • 26 Jul 2019 • Kartikeya Bhardwaj, Chingyi Lin, Anderson Sartor, Radu Marculescu
Therefore, we propose Network of Neural Networks (NoNN), a new distributed IoT learning paradigm that compresses a large pretrained 'teacher' deep network into several disjoint and highly-compressed 'student' modules, without loss of accuracy.
no code implementations • 17 May 2019 • Kartikeya Bhardwaj, Naveen Suda, Radu Marculescu
Model compression is eminently suited for deploying deep learning on IoT-devices.
no code implementations • 20 Jan 2019 • Brian Davis, Umang Bhatt, Kartikeya Bhardwaj, Radu Marculescu, José M. F. Moura
In this paper, we present a new approach to interpret deep learning models.
no code implementations • 1 Dec 2018 • Jiqian Dong, Gopaljee Atulya, Kartikeya Bhardwaj, Radu Marculescu
To this end, we propose a new network science- and representation learning-based approach that can quantify economic indicators and visualize the growth of various regions.