Search Results for author: Baolin Li

Found 10 papers, 0 papers with code

Toward Sustainable GenAI using Generation Directives for Carbon-Friendly Large Language Model Inference

no code implementations19 Mar 2024 Baolin Li, Yankai Jiang, Vijay Gadepally, Devesh Tiwari

The rapid advancement of Generative Artificial Intelligence (GenAI) across diverse sectors raises significant environmental concerns, notably the carbon emissions from their cloud and high performance computing (HPC) infrastructure.

Language Modelling Large Language Model

Sustainable Supercomputing for AI: GPU Power Capping at HPC Scale

no code implementations25 Feb 2024 Dan Zhao, Siddharth Samsi, Joseph McDonald, Baolin Li, David Bestor, Michael Jones, Devesh Tiwari, Vijay Gadepally

In this paper, we study the aggregate effect of power-capping GPUs on GPU temperature and power draw at a research supercomputing center.

KAIROS: Building Cost-Efficient Machine Learning Inference Systems with Heterogeneous Cloud Resources

no code implementations12 Oct 2022 Baolin Li, Siddharth Samsi, Vijay Gadepally, Devesh Tiwari

Online inference is becoming a key service product for many businesses, deployed in cloud platforms to meet customer demands.

Great Power, Great Responsibility: Recommendations for Reducing Energy for Training Language Models

no code implementations Findings (NAACL) 2022 Joseph McDonald, Baolin Li, Nathan Frey, Devesh Tiwari, Vijay Gadepally, Siddharth Samsi

In particular, we focus on techniques to measure energy usage and different hardware and datacenter-oriented settings that can be tuned to reduce energy consumption for training and inference for language models.

Cloud Computing Language Modelling

Benchmarking Resource Usage for Efficient Distributed Deep Learning

no code implementations28 Jan 2022 Nathan C. Frey, Baolin Li, Joseph McDonald, Dan Zhao, Michael Jones, David Bestor, Devesh Tiwari, Vijay Gadepally, Siddharth Samsi

Deep learning (DL) workflows demand an ever-increasing budget of compute and energy in order to achieve outsized gains.

Benchmarking

Cannot find the paper you are looking for? You can Submit a new open access paper.