no code implementations • 2 Feb 2024 • Guangfeng Yan, Tan Li, Yuanzhang Xiao, Hanxu Hou, Linqi Song
We consider a general family of heavy-tail gradients that follow a power-law distribution, we aim to minimize the error resulting from quantization, thereby determining optimal values for two critical parameters: the truncation threshold and the quantization density.
no code implementations • 2 Feb 2024 • Guangfeng Yan, Tan Li, Yuanzhang Xiao, Congduan Li, Linqi Song
To address the communication bottleneck challenge in distributed learning, our work introduces a novel two-stage quantization strategy designed to enhance the communication efficiency of distributed Stochastic Gradient Descent (SGD).
no code implementations • 26 Apr 2023 • Guangfeng Yan, Tan Li, Kui Wu, Linqi Song
Communication efficiency and privacy protection are two critical issues in distributed machine learning.
no code implementations • 24 Oct 2022 • Mengzhe Ruan, Guangfeng Yan, Yuanzhang Xiao, Linqi Song, Weitao Xu
This paper proposes a novel adaptive Top-K in SGD framework that enables an adaptive degree of sparsification for each gradient descent step to optimize the convergence performance by balancing the trade-off between communication cost and convergence error.
no code implementations • 30 Jul 2021 • Guangfeng Yan, Shao-Lun Huang, Tian Lan, Linqi Song
Gradient quantization is an emerging technique in reducing communication costs in distributed learning.
no code implementations • 1 Jan 2021 • Guangfeng Yan, Shao-Lun Huang, Tian Lan, Linqi Song
This paper addresses this issue by proposing a novel dynamic quantized SGD (DQSGD) framework, which enables us to optimize the quantization strategy for each gradient descent step by exploring the trade-off between communication cost and modeling error.
1 code implementation • ACL 2020 • Guangfeng Yan, Lu Fan, Qimai Li, Han Liu, Xiaotong Zhang, Xiao-Ming Wu, Albert Y. S. Lam
User intent classification plays a vital role in dialogue systems.