Search Results for author: Zeman Li

Found 3 papers, 1 papers with code

Synthetic Text Generation for Training Large Language Models via Gradient Matching

no code implementations24 Feb 2025 Dang Nguyen, Zeman Li, Mohammadhossein Bateni, Vahab Mirrokni, Meisam Razaviyayn, Baharan Mirzasoleiman

In this work, we propose the first theoretically rigorous approach for generating synthetic human-readable text that guarantees the convergence and performance of LLMs during fine-tuning on a target task.

Text Generation

Addax: Utilizing Zeroth-Order Gradients to Improve Memory Efficiency and Performance of SGD for Fine-Tuning Language Models

no code implementations9 Oct 2024 Zeman Li, Xinwei Zhang, Peilin Zhong, Yuan Deng, Meisam Razaviyayn, Vahab Mirrokni

In our experiments on the larger OPT-30B model, on average, Addax outperforms MeZO in terms of accuracy/F1 score by >16 and runs 30x faster on a single H100 GPU.

Optimal Differentially Private Model Training with Public Data

1 code implementation26 Jun 2023 Andrew Lowy, Zeman Li, Tianjian Huang, Meisam Razaviyayn

We show that the optimal error rates can be attained (up to log factors) by either discarding private data and training a public model, or treating public data like it is private and using an optimal DP algorithm.

model

Cannot find the paper you are looking for? You can Submit a new open access paper.