Search Results for author: Zaoxing Liu

Found 3 papers, 1 papers with code

GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM

1 code implementation8 Mar 2024 Hao Kang, Qingru Zhang, Souvik Kundu, Geonhwa Jeong, Zaoxing Liu, Tushar Krishna, Tuo Zhao

Key-value (KV) caching has become the de-facto to accelerate generation speed for large language models (LLMs) inference.

Quantization

Enhancing the Privacy of Federated Learning with Sketching

no code implementations5 Nov 2019 Zaoxing Liu, Tian Li, Virginia Smith, Vyas Sekar

Federated learning methods run training tasks directly on user devices and do not share the raw user data with third parties.

Federated Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.