Search Results for author: Ana Klimovic

Found 8 papers, 4 papers with code

On Distributed Larger-Than-Memory Subset Selection With Pairwise Submodular Functions

no code implementations26 Feb 2024 Maximilian Böther, Abraham Sebastian, Pranjal Awasthi, Ana Klimovic, Srikumar Ramalingam

In this paper, we relax the requirement of having a central machine for the target subset by proposing a novel distributed bounding algorithm with provable approximation guarantees.

Modyn: A Platform for Model Training on Dynamic Datasets With Sample-Level Data Selection

no code implementations11 Dec 2023 Maximilian Böther, Viktor Gsteiger, Ties Robroek, Ana Klimovic

Machine learning training data is often dynamic in real-world use cases, i. e., data is added or removed and may experience distribution shifts over time.

Recommendation Systems

DeltaZip: Multi-Tenant Language Model Serving via Delta Compression

1 code implementation8 Dec 2023 Xiaozhe Yao, Ana Klimovic

Fine-tuning large language models (LLMs) for downstream tasks can greatly improve model quality, however serving many different fine-tuned LLMs concurrently for users in multi-tenant environments is challenging.

Language Modelling

tf.data service: A Case for Disaggregating ML Input Data Processing

no code implementations26 Oct 2022 Andrew Audibert, Yang Chen, Dan Graur, Ana Klimovic, Jiri Simsa, Chandramohan A. Thekkath

To avoid data stalls, the host CPU and RAM required for input data processing per accelerator core used for ML computations varies across jobs.

SHiFT: An Efficient, Flexible Search Engine for Transfer Learning

1 code implementation4 Apr 2022 Cedric Renggli, Xiaozhe Yao, Luka Kolar, Luka Rimanic, Ana Klimovic, Ce Zhang

Transfer learning can be seen as a data- and compute-efficient alternative to training models from scratch.

Transfer Learning

Plumber: Diagnosing and Removing Performance Bottlenecks in Machine Learning Data Pipelines

2 code implementations7 Nov 2021 Michael Kuchnik, Ana Klimovic, Jiri Simsa, Virginia Smith, George Amvrosiadis

Our analysis of over two million ML jobs in Google datacenters reveals that a significant fraction of model training jobs could benefit from faster input data pipelines.

BIG-bench Machine Learning

Towards Demystifying Serverless Machine Learning Training

1 code implementation17 May 2021 Jiawei Jiang, Shaoduo Gan, Yue Liu, Fanlin Wang, Gustavo Alonso, Ana Klimovic, Ankit Singla, Wentao Wu, Ce Zhang

The appeal of serverless (FaaS) has triggered a growing interest on how to use it in data-intensive applications such as ETL, query processing, or machine learning (ML).

BIG-bench Machine Learning

tf.data: A Machine Learning Data Processing Framework

no code implementations28 Jan 2021 Derek G. Murray, Jiri Simsa, Ana Klimovic, Ihor Indyk

Finally, we characterize machine learning input pipelines for millions of jobs that ran in Google's fleet, showing that input data processing is highly diverse and consumes a significant fraction of job resources.

BIG-bench Machine Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.