Search Results for author: Mosharaf Chowdhury

Found 25 papers, 15 papers with code

EXP-Bench: Can AI Conduct AI Research Experiments?

1 code implementation30 May 2025 Patrick Tser Jern Kon, Jiachen Liu, Xinyi Zhu, Qiuyi Ding, Jingjia Peng, Jiarong Xing, Yibo Huang, Yiming Qiu, Jayanth Srinivasa, Myungjin Lee, Mosharaf Chowdhury, Matei Zaharia, Ang Chen

To enable the creation of such intricate and authentic tasks with high-fidelity, we design a semi-autonomous pipeline to extract and structure crucial experimental details from these research papers and their associated open-source code.

The ML.ENERGY Benchmark: Toward Automated Inference Energy Measurement and Optimization

3 code implementations9 May 2025 Jae-Won Chung, Jiachen Liu, Jeff J. Ma, Ruofan Wu, Oh Jun Kweon, Yuxuan Xia, Zhiyu Wu, Mosharaf Chowdhury

As the adoption of Generative AI in real-world services grow explosively, energy has emerged as a critical bottleneck resource.

Benchmarking

Curie: Toward Rigorous and Automated Scientific Experimentation with AI Agents

1 code implementation22 Feb 2025 Patrick Tser Jern Kon, Jiachen Liu, Qiuyi Ding, Yiming Qiu, Zhenning Yang, Yibo Huang, Jayanth Srinivasa, Myungjin Lee, Mosharaf Chowdhury, Ang Chen

Scientific experimentation, a cornerstone of human progress, demands rigor in reliability, methodical control, and interpretability to yield meaningful results.

AI Agent

Andes: Defining and Enhancing Quality-of-Experience in LLM-Based Text Streaming Services

1 code implementation25 Apr 2024 Jiachen Liu, Jae-Won Chung, Zhiyu Wu, Fan Lai, Myungjin Lee, Mosharaf Chowdhury

Large language models (LLMs) are now at the core of conversational AI services such as real-time translation and chatbots, which provide live user interaction by incrementally streaming text to the user.

FedTrans: Efficient Federated Learning via Multi-Model Transformation

no code implementations21 Apr 2024 Yuxuan Zhu, Jiachen Liu, Mosharaf Chowdhury, Fan Lai

Federated learning (FL) aims to train machine learning (ML) models across potentially millions of edge client devices.

Federated Learning

Toward Cross-Layer Energy Optimizations in AI Systems

no code implementations10 Apr 2024 Jae-Won Chung, Nishil Talati, Mosharaf Chowdhury

The "AI for Science, Energy, and Security" report from DOE outlines a significant focus on developing and optimizing artificial intelligence workflows for a foundational impact on a broad range of DOE missions.

Language Modeling Language Modelling +1

Venn: Resource Management for Collaborative Learning Jobs

1 code implementation13 Dec 2023 Jiachen Liu, Fan Lai, Ding Ding, Yiwen Zhang, Mosharaf Chowdhury

In this paper, we present Venn, a CL resource manager that efficiently schedules ephemeral, heterogeneous devices among multiple CL jobs to reduce the average job completion time (JCT).

Federated Learning Management +1

Reducing Energy Bloat in Large Model Training

2 code implementations12 Dec 2023 Jae-Won Chung, Yile Gu, Insu Jang, Luoxi Meng, Nikhil Bansal, Mosharaf Chowdhury

Training large AI models on numerous GPUs consumes a massive amount of energy, making power delivery one of the largest limiting factors in building and operating datacenters for AI workloads.

model

Efficient Large Language Models: A Survey

3 code implementations6 Dec 2023 Zhongwei Wan, Xin Wang, Che Liu, Samiul Alam, Yu Zheng, Jiachen Liu, Zhongnan Qu, Shen Yan, Yi Zhu, Quanlu Zhang, Mosharaf Chowdhury, Mi Zhang

We hope our survey can serve as a valuable resource to help researchers and practitioners gain a systematic understanding of efficient LLMs research and inspire them to contribute to this important and exciting field.

Natural Language Understanding Survey +1

Oobleck: Resilient Distributed Training of Large Models Using Pipeline Templates

2 code implementations15 Sep 2023 Insu Jang, Zhenning Yang, Zhen Zhang, Xin Jin, Mosharaf Chowdhury

Oobleck enables resilient distributed training of large DNN models with guaranteed fault tolerance.

Chasing Low-Carbon Electricity for Practical and Sustainable DNN Training

1 code implementation4 Mar 2023 Zhenning Yang, Luoxi Meng, Jae-Won Chung, Mosharaf Chowdhury

Specifically, our solution observes real-time carbon intensity shifts during training and controls the energy consumption of GPUs, thereby reducing carbon footprint while maintaining training performance.

FLINT: A Platform for Federated Learning Integration

no code implementations24 Feb 2023 Ewen Wang, Ajay Kannan, Yuefeng Liang, Boyi Chen, Mosharaf Chowdhury

Cross-device federated learning (FL) has been well-studied from algorithmic, system scalability, and training speed perspectives.

Federated Learning

DPack: Efficiency-Oriented Privacy Budget Scheduling

2 code implementations26 Dec 2022 Pierre Tholoniat, Kelly Kostopoulou, Mosharaf Chowdhury, Asaf Cidon, Roxana Geambasu, Mathias Lécuyer, Junfeng Yang

This DP budget can be regarded as a new type of compute resource in workloads of multiple ML models training on user data.

Fairness Scheduling

Auxo: Efficient Federated Learning via Scalable Client Clustering

no code implementations29 Oct 2022 Jiachen Liu, Fan Lai, Yinwei Dai, Aditya Akella, Harsha Madhyastha, Mosharaf Chowdhury

In this paper, we explore an additional layer of complexity to mitigate such heterogeneity by grouping clients with statistically similar data distributions (cohorts).

Clustering Federated Learning

Zeus: Understanding and Optimizing GPU Energy Consumption of DNN Training

1 code implementation12 Aug 2022 Jie You, Jae-Won Chung, Mosharaf Chowdhury

In this paper, we observe that common practices to improve training performance can often lead to inefficient energy usage.

Swan: A Neural Engine for Efficient DNN Training on Smartphone SoCs

no code implementations9 Jun 2022 Sanjay Sri Vallabh Singapuram, Fan Lai, Chuheng Hu, Mosharaf Chowdhury

The need to train DNN models on end-user devices (e. g., smartphones) is increasing with the need to improve data privacy and reduce communication overheads.

Egeria: Efficient DNN Training with Knowledge-Guided Layer Freezing

no code implementations17 Jan 2022 Yiding Wang, Decang Sun, Kai Chen, Fan Lai, Mosharaf Chowdhury

To explore this, we first introduce the notion of training plasticity to quantify the training progress of internal DNN layers.

Quantization

Treehouse: A Case For Carbon-Aware Datacenter Software

no code implementations6 Jan 2022 Thomas Anderson, Adam Belay, Mosharaf Chowdhury, Asaf Cidon, Irene Zhang

The end of Dennard scaling and the slowing of Moore's Law has put the energy use of datacenters on an unsustainable path.

Fed-ensemble: Improving Generalization through Model Ensembling in Federated Learning

1 code implementation21 Jul 2021 Naichen Shi, Fan Lai, Raed Al Kontar, Mosharaf Chowdhury

In this paper we propose Fed-ensemble: a simple approach that bringsmodel ensembling to federated learning (FL).

Federated Learning

FedScale: Benchmarking Model and System Performance of Federated Learning at Scale

2 code implementations24 May 2021 Fan Lai, Yinwei Dai, Sanjay S. Singapuram, Jiachen Liu, Xiangfeng Zhu, Harsha V. Madhyastha, Mosharaf Chowdhury

We present FedScale, a federated learning (FL) benchmarking suite with realistic datasets and a scalable runtime to enable reproducible FL research.

Benchmarking Federated Learning +8

Oort: Efficient Federated Learning via Guided Participant Selection

1 code implementation12 Oct 2020 Fan Lai, Xiangfeng Zhu, Harsha V. Madhyastha, Mosharaf Chowdhury

In this paper, we propose Oort to improve the performance of federated training and testing with guided participant selection.

Federated Learning

Salus: Fine-Grained GPU Sharing Primitives for Deep Learning Applications

1 code implementation12 Feb 2019 Peifeng Yu, Mosharaf Chowdhury

We show that these primitives can then be used to implement flexible sharing policies such as fairness, prioritization, and packing for various use cases.

Deep Learning Fairness +2

Fast and Accurate Performance Analysis of LTE Radio Access Networks

no code implementations16 May 2016 Anand Padmanabha Iyer, Ion Stoica, Mosharaf Chowdhury, Li Erran Li

Our choice of this domain is influenced by its commonalities with several other domains that produce real-time data, our access to a large live dataset, and their real-time nature and dimensionality which makes it a natural fit for a popular analysis technique, machine learning (ML).

Feature Engineering Multi-Task Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.