Search Results for author: Greg Ganger

Found 2 papers, 1 papers with code

PipeDream: Fast and Efficient Pipeline Parallel DNN Training

1 code implementation8 Jun 2018 Aaron Harlap, Deepak Narayanan, Amar Phanishayee, Vivek Seshadri, Nikhil Devanur, Greg Ganger, Phil Gibbons

PipeDream is a Deep Neural Network(DNN) training system for GPUs that parallelizes computation by pipelining execution across multiple machines.

Distributed, Parallel, and Cluster Computing

More Effective Distributed ML via a Stale Synchronous Parallel Parameter Server

no code implementations NeurIPS 2013 Qirong Ho, James Cipar, Henggang Cui, Seunghak Lee, Jin Kyu Kim, Phillip B. Gibbons, Garth A. Gibson, Greg Ganger, Eric P. Xing

We propose a parameter server system for distributed ML, which follows a Stale Synchronous Parallel (SSP) model of computation that maximizes the time computational workers spend doing useful work on ML algorithms, while still providing correctness guarantees.

Cannot find the paper you are looking for? You can Submit a new open access paper.