Search Results for author: Saurabh Agarwal

Found 14 papers, 6 papers with code

Decoding Speculative Decoding

no code implementations2 Feb 2024 Minghao Yan, Saurabh Agarwal, Shivaram Venkataraman

However, our experiments indicate the contrary with throughput diminishing as the probability of generated tokens to be accepted by the target model increases.

MultiFusionNet: Multilayer Multimodal Fusion of Deep Neural Networks for Chest X-Ray Image Classification

no code implementations1 Jan 2024 Saurabh Agarwal, K. V. Arya, Yogesh Kumar Meena

The proposed multilayer multimodal fusion model, along with the FDSFM module, holds promise for accurate disease classification and can also be extended to other disease classifications in chest X-ray images.

Image Classification

Cuttlefish: Low-Rank Model Training without All the Tuning

1 code implementation4 May 2023 Hongyi Wang, Saurabh Agarwal, Pongsakorn U-chupala, Yoshiki Tanaka, Eric P. Xing, Dimitris Papailiopoulos

Cuttlefish leverages the observation that after a few epochs of full-rank training, the stable rank (i. e., an approximation of the true rank) of each layer stabilizes at a constant value.

BagPipe: Accelerating Deep Recommendation Model Training

no code implementations24 Feb 2022 Saurabh Agarwal, Chengpo Yan, Ziyi Zhang, Shivaram Venkataraman

Based on these insights, we develop Bagpipe, a system for training deep recommendation models that uses caching and prefetching to overlap remote embedding accesses with the computation.

Pufferfish: Communication-efficient Models At No Extra Cost

1 code implementation5 Mar 2021 Hongyi Wang, Saurabh Agarwal, Dimitris Papailiopoulos

In this work, we present Pufferfish, a communication and computation efficient distributed training framework that incorporates the gradient compression into the model training process via training low-rank, pre-factorized deep networks.

Quantization

On the Utility of Gradient Compression in Distributed Training Systems

1 code implementation28 Feb 2021 Saurabh Agarwal, Hongyi Wang, Shivaram Venkataraman, Dimitris Papailiopoulos

A rich body of prior work has highlighted the existence of communication bottlenecks in synchronous data-parallel training.

Model Compression

AutoFreeze: Automatically Freezing Model Blocks to Accelerate Fine-tuning

1 code implementation2 Feb 2021 YuHan Liu, Saurabh Agarwal, Shivaram Venkataraman

With the rapid adoption of machine learning (ML), a number of domains now use the approach of fine tuning models which were pre-trained on a large corpus of data.

Accordion: Adaptive Gradient Communication via Critical Learning Regime Identification

3 code implementations29 Oct 2020 Saurabh Agarwal, Hongyi Wang, Kangwook Lee, Shivaram Venkataraman, Dimitris Papailiopoulos

The techniques usually require choosing a static compression ratio, often requiring users to balance the trade-off between model accuracy and per-iteration speedup.

Quantization

Scalable K-Medoids via True Error Bound and Familywise Bandits

no code implementations27 May 2019 Aravindakshan Babu, Saurabh Agarwal, Sudarshan Babu, Hariharan Chandrasekaran

K-Medoids(KM) is a standard clustering method, used extensively on semi-metric data. Error analyses of KM have traditionally used an in-sample notion of error, which can be far from the true error and suffer from generalization gap.

Clustering

Graph based Question Answering System

no code implementations5 Dec 2018 Piyush Mital, Saurabh Agarwal, Bhargavi Neti, Yashodhara Haribhakta, Vibhavari Kamble, Krishnanjan Bhattacharjee, Debashri Das, Swati Mehta, Ajai Kumar

In today's digital age in the dawning era of big data analytics it is not the information but the linking of information through entities and actions which defines the discourse.

Question Answering Retrieval

Cannot find the paper you are looking for? You can Submit a new open access paper.