Search Results for author: Xiaozhe Yao

Found 7 papers, 5 papers with code

Mixtera: A Data Plane for Foundation Model Training

1 code implementation27 Feb 2025 Maximilian Böther, Xiaozhe Yao, Tolga Kerimoglu, Dan Graur, Viktor Gsteiger, Ana Klimovic

We build and present Mixtera, a data plane for foundation model training that enables users to declaratively express which data samples should be used in which proportion and in which order during training.

model

DeltaZip: Efficient Serving of Multiple Full-Model-Tuned LLMs

1 code implementation8 Dec 2023 Xiaozhe Yao, Qinghao Hu, Ana Klimovic

Fine-tuning large language models (LLMs) greatly improves model quality for downstream tasks.

Language Modelling

SHiFT: An Efficient, Flexible Search Engine for Transfer Learning

1 code implementation4 Apr 2022 Cedric Renggli, Xiaozhe Yao, Luka Kolar, Luka Rimanic, Ana Klimovic, Ce Zhang

Transfer learning can be seen as a data- and compute-efficient alternative to training models from scratch.

Transfer Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.