Search Results for author: Radu Marculescu

Found 36 papers, 20 papers with code

RapidNet: Multi-Level Dilated Convolution Based Mobile Backbone

1 code implementation14 Dec 2024 Mustafa Munir, Md Mostafijur Rahman, Radu Marculescu

Experiments show that our proposed model outperforms state-of-the-art (SOTA) mobile CNN, ViT, ViG, and hybrid architectures in terms of accuracy and/or speed on image classification, object detection, instance segmentation, and semantic segmentation.

Image Classification Instance Segmentation +3

Online-LoRA: Task-free Online Continual Learning via Low Rank Adaptation

1 code implementation8 Nov 2024 Xiwen Wei, Guihong Li, Radu Marculescu

Catastrophic forgetting is a significant challenge in online continual learning (OCL), especially for non-stationary data streams that do not have well-defined task boundaries.

Continual Learning

Skip2-LoRA: A Lightweight On-device DNN Fine-tuning Method for Low-cost Edge Devices

no code implementations28 Oct 2024 Hiroki Matsutani, Masaaki Kondo, Kazuki Sunaga, Radu Marculescu

This paper proposes Skip2-LoRA as a lightweight fine-tuning method for deep neural networks to address the gap between pre-trained and deployed models.

CrediRAG: Network-Augmented Credibility-Based Retrieval for Misinformation Detection in Reddit

no code implementations15 Oct 2024 Ashwin Ram, Yigit Ege Bayiz, Arash Amini, Mustafa Munir, Radu Marculescu

Fake news threatens democracy and exacerbates the polarization and divisions in society; therefore, accurately detecting online misinformation is the foundation of addressing this issue.

Fake News Detection Misinformation +1

A Tiny Supervised ODL Core with Auto Data Pruning for Human Activity Recognition

no code implementations2 Aug 2024 Hiroki Matsutani, Radu Marculescu

As a tinyML solution at a few mW for the human activity recognition, we design a supervised ODL core that supports our automatic data pruning using a 45nm CMOS process technology.

Human Activity Recognition

Scaling Graph Convolutions for Mobile Vision

1 code implementation9 Jun 2024 William Avery, Mustafa Munir, Radu Marculescu

To compete with existing mobile architectures, MobileViG introduces Sparse Vision Graph Attention (SVGA), a fast token-mixing operator based on the principles of GNNs.

Graph Attention Graph Neural Network +5

Ada-VE: Training-Free Consistent Video Editing Using Adaptive Motion Prior

1 code implementation7 Jun 2024 Tanvir Mahmud, Mustafa Munir, Radu Marculescu, Diana Marculescu

This enables a greater number of cross-frame attentions over more frames within the same computational budget, thereby enhancing both video quality and temporal coherence.

Consistent Character Generation Optical Flow Estimation +2

PP-SAM: Perturbed Prompts for Robust Adaptation of Segment Anything Model for Polyp Segmentation

1 code implementation27 May 2024 Md Mostafijur Rahman, Mustafa Munir, Debesh Jha, Ulas Bagci, Radu Marculescu

To this end, we utilize variable perturbed bounding box prompts (BBP) to enrich the learning context and enhance the model's robustness to BBP perturbations during inference.

Segmentation

GreedyViG: Dynamic Axial Graph Construction for Efficient Vision GNNs

1 code implementation CVPR 2024 Mustafa Munir, William Avery, Md Mostafijur Rahman, Radu Marculescu

Our smallest model, GreedyViG-S, achieves 81. 1% top-1 accuracy on ImageNet-1K, 2. 9% higher than Vision GNN and 2. 2% higher than Vision HyperGraph Neural Network (ViHGNN), with less GMACs and a similar number of parameters.

graph construction Image Classification +4

News Source Credibility Assessment: A Reddit Case Study

no code implementations7 Feb 2024 Arash Amini, Yigit Ege Bayiz, Ashwin Ram, Radu Marculescu, Ufuk Topcu

In the era of social media platforms, identifying the credibility of online content is crucial to combat misinformation.

Binary Classification Misinformation

Machine Unlearning for Image-to-Image Generative Models

2 code implementations1 Feb 2024 Guihong Li, Hsiang Hsu, Chun-Fu Chen, Radu Marculescu

This paper serves as a bridge, addressing the gap by providing a unifying framework of machine unlearning for image-to-image generative models.

Machine Unlearning

Fast-NTK: Parameter-Efficient Unlearning for Large-Scale Models

no code implementations22 Dec 2023 Guihong Li, Hsiang Hsu, Chun-Fu Chen, Radu Marculescu

The rapid growth of machine learning has spurred legislative initiatives such as ``the Right to be Forgotten,'' allowing users to request data removal.

Machine Unlearning parameter-efficient fine-tuning

G-CASCADE: Efficient Cascaded Graph Convolutional Decoding for 2D Medical Image Segmentation

1 code implementation24 Oct 2023 Md Mostafijur Rahman, Radu Marculescu

The encoder utilizes the self-attention mechanism to capture long-range dependencies, while the decoder refines the feature maps preserving long-range information due to the global receptive fields of the graph convolution block.

 Ranked #1 on Retinal Vessel Segmentation on DRIVE (Specificity metric)

Decoder Image Segmentation +3

Zero-Shot Neural Architecture Search: Challenges, Solutions, and Opportunities

1 code implementation5 Jul 2023 Guihong Li, Duc Hoang, Kartikeya Bhardwaj, Ming Lin, Zhangyang Wang, Radu Marculescu

Recently, zero-shot (or training-free) Neural Architecture Search (NAS) approaches have been proposed to liberate NAS from the expensive training process.

Neural Architecture Search

MobileViG: Graph-Based Sparse Attention for Mobile Vision Applications

1 code implementation1 Jul 2023 Mustafa Munir, William Avery, Radu Marculescu

Our work proves that well designed hybrid CNN-GNN architectures can be a new avenue of exploration for designing models that are extremely fast and accurate on mobile devices.

Graph Attention Image Classification +4

TIPS: Topologically Important Path Sampling for Anytime Neural Networks

no code implementations13 May 2023 Guihong Li, Kartikeya Bhardwaj, Yuedong Yang, Radu Marculescu

Anytime neural networks (AnytimeNNs) are a promising solution to adaptively adjust the model complexity at runtime under various hardware resource constraints.

ZiCo: Zero-shot NAS via Inverse Coefficient of Variation on Gradients

1 code implementation26 Jan 2023 Guihong Li, Yuedong Yang, Kartikeya Bhardwaj, Radu Marculescu

Based on this theoretical analysis, we propose a new zero-shot proxy, ZiCo, the first proxy that works consistently better than #Params.

Image Classification Neural Architecture Search

Medical Image Segmentation via Cascaded Attention Decoding

1 code implementation Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) 2023 Md Mostafijur Rahman, Radu Marculescu

To address this issue, we propose a novel attention-based decoder, namely CASCaded Attention DEcoder (CASCADE), which leverages the multiscale features of hierarchical vision transformers.

Decoder Image Segmentation +2

Efficient On-device Training via Gradient Filtering

1 code implementation CVPR 2023 Yuedong Yang, Guihong Li, Radu Marculescu

Despite its importance for federated learning, continuous learning and many other applications, on-device training remains an open problem for EdgeAI.

Federated Learning Image Classification +1

Dynamic Multimodal Fusion

1 code implementation31 Mar 2022 Zihui Xue, Radu Marculescu

In this work, we propose dynamic multimodal fusion (DynMM), a new approach that adaptively fuses multimodal data and generates data-dependent forward paths during inference.

Computational Efficiency Semantic Segmentation +1

SUGAR: Efficient Subgraph-level Training via Resource-aware Graph Partitioning

no code implementations31 Jan 2022 Zihui Xue, Yuedong Yang, Mengtian Yang, Radu Marculescu

Graph Neural Networks (GNNs) have demonstrated a great potential in a variety of graph-based applications, such as recommender systems, drug discovery, and object recognition.

Drug Discovery Edge-computing +3

FLASH: Fast Neural Architecture Search with Hardware Optimization

no code implementations1 Aug 2021 Guihong Li, Sumit K. Mandal, Umit Y. Ogras, Radu Marculescu

This paper proposes FLASH, a very fast NAS methodology that co-optimizes the DNN accuracy and performance on a real hardware platform.

Neural Architecture Search

How does topology of neural architectures impact gradient propagation and model performance?

1 code implementation CVPR 2021 Kartikeya Bhardwa, Guihong Li2, Radu Marculescu

In this paper, we reveal that the topology of the concatenation-type skip connections is closely related to the gradient propagation which, in turn, enables a predictable behavior of DNNs’ test performance.

Model Compression Neural Architecture Search

On the relationship between topology and gradient propagation in deep networks

no code implementations1 Jan 2021 Kartikeya Bhardwaj, Guihong Li, Radu Marculescu

(ii) Can certain topological characteristics of deep networks indicate a priori (i. e., without training) which models, with a different number of parameters/FLOPS/layers, achieve a similar accuracy?

New Directions in Distributed Deep Learning: Bringing the Network at Forefront of IoT Design

no code implementations25 Aug 2020 Kartikeya Bhardwaj, Wei Chen, Radu Marculescu

In this paper, we first highlight three major challenges to large-scale adoption of deep learning at the edge: (i) Hardware-constrained IoT devices, (ii) Data security and privacy in the IoT era, and (iii) Lack of network-aware deep learning algorithms for distributed inference across multiple IoT devices.

Deep Learning Federated Learning

FedMAX: Mitigating Activation Divergence for Accurate and Communication-Efficient Federated Learning

1 code implementation7 Apr 2020 Wei Chen, Kartikeya Bhardwaj, Radu Marculescu

In this paper, we identify a new phenomenon called activation-divergence which occurs in Federated Learning (FL) due to data heterogeneity (i. e., data being non-IID) across multiple users.

Federated Learning

EdgeAI: A Vision for Deep Learning in IoT Era

no code implementations23 Oct 2019 Kartikeya Bhardwaj, Naveen Suda, Radu Marculescu

The significant computational requirements of deep learning present a major bottleneck for its large-scale adoption on hardware-constrained IoT-devices.

Deep Learning

How does topology influence gradient propagation and model performance of deep networks with DenseNet-type skip connections?

2 code implementations CVPR 2021 Kartikeya Bhardwaj, Guihong Li, Radu Marculescu

In this paper, we reveal that the topology of the concatenation-type skip connections is closely related to the gradient propagation which, in turn, enables a predictable behavior of DNNs' test performance.

Model Compression

Memory- and Communication-Aware Model Compression for Distributed Deep Learning Inference on IoT

no code implementations26 Jul 2019 Kartikeya Bhardwaj, Chingyi Lin, Anderson Sartor, Radu Marculescu

Therefore, we propose Network of Neural Networks (NoNN), a new distributed IoT learning paradigm that compresses a large pretrained 'teacher' deep network into several disjoint and highly-compressed 'student' modules, without loss of accuracy.

Image Classification Model Compression

A Dynamic Network and Representation LearningApproach for Quantifying Economic Growth fromSatellite Imagery

no code implementations1 Dec 2018 Jiqian Dong, Gopaljee Atulya, Kartikeya Bhardwaj, Radu Marculescu

To this end, we propose a new network science- and representation learning-based approach that can quantify economic indicators and visualize the growth of various regions.

Representation Learning

Learning-based Application-Agnostic 3D NoC Design for Heterogeneous Manycore Systems

1 code implementation20 Oct 2018 Biresh Kumar Joardar, Ryan Gary Kim, Janardhan Rao Doppa, Partha Pratim Pande, Diana Marculescu, Radu Marculescu

Our results show that these generalized 3D NoCs only incur a 1. 8% (36-tile system) and 1. 1% (64-tile system) average performance loss compared to application-specific NoCs.

Machine Learning and Manycore Systems Design: A Serendipitous Symbiosis

no code implementations30 Nov 2017 Ryan Gary Kim, Janardhan Rao Doppa, Partha Pratim Pande, Diana Marculescu, Radu Marculescu

Tight collaboration between experts of machine learning and manycore system design is necessary to create a data-driven manycore design framework that integrates both learning and expert knowledge.

BIG-bench Machine Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.