Search Results for author: Keshi Ge

Found 3 papers, 1 papers with code

Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models

1 code implementation10 Jun 2022 Zhiquan Lai, Shengwei Li, Xudong Tang, Keshi Ge, Weijie Liu, Yabo Duan, Linbo Qiao, Dongsheng Li

These features make it necessary to apply 3D parallelism, which integrates data parallelism, pipeline model parallelism and tensor model parallelism, to achieve high training efficiency.

S2 Reducer: High-Performance Sparse Communication to Accelerate Distributed Deep Learning

no code implementations5 Oct 2021 Keshi Ge, Yongquan Fu, Zhiquan Lai, Xiaoge Deng, Dongsheng Li

Distributed stochastic gradient descent (SGD) approach has been widely used in large-scale deep learning, and the gradient collective method is vital to ensure the training scalability of the distributed deep learning system.

Vocal Bursts Intensity Prediction

An Efficient ADMM-Based Algorithm to Nonconvex Penalized Support Vector Machines

no code implementations11 Sep 2018 Lei Guan, Linbo Qiao, Dongsheng Li, Tao Sun, Keshi Ge, Xicheng Lu

Support vector machines (SVMs) with sparsity-inducing nonconvex penalties have received considerable attentions for the characteristics of automatic classification and variable selection.

General Classification Variable Selection

Cannot find the paper you are looking for? You can Submit a new open access paper.