Search Results for author: Dimitrios Vytiniotis

Found 10 papers, 3 papers with code

Automatic Discovery of Composite SPMD Partitioning Strategies in PartIR

no code implementations7 Oct 2022 Sami Alabed, Dominik Grewe, Juliana Franco, Bart Chrzaszcz, Tom Natan, Tamara Norman, Norman A. Rink, Dimitrios Vytiniotis, Michael Schaarschmidt

Large neural network models are commonly trained through a combination of advanced parallelism strategies in a single program, multiple data (SPMD) paradigm.

Automap: Towards Ergonomic Automated Parallelism for ML Models

no code implementations6 Dec 2021 Michael Schaarschmidt, Dominik Grewe, Dimitrios Vytiniotis, Adam Paszke, Georg Stefan Schmid, Tamara Norman, James Molloy, Jonathan Godwin, Norman Alexander Rink, Vinod Nair, Dan Belov

The rapid rise in demand for training large neural network architectures has brought into focus the need for partitioning strategies, for example by using data, model, or pipeline parallelism.

Memory-efficient array redistribution through portable collective communication

no code implementations2 Dec 2021 Norman A. Rink, Adam Paszke, Dimitrios Vytiniotis, Georg Stefan Schmid

In this paper we address the problem of redistributing multi-dimensional array data in SPMD computations, the most prevalent form of parallelism in deep learning.

Synthesizing Optimal Parallelism Placement and Reduction Strategies on Hierarchical Systems for Deep Learning

no code implementations20 Oct 2021 Ningning Xie, Tamara Norman, Dominik Grewe, Dimitrios Vytiniotis

We present a novel characterization of the mapping of multiple parallelism forms (e. g. data and model parallelism) onto hierarchical accelerator systems that is hierarchy-aware and greatly reduces the space of software-to-hardware mapping.

Program Synthesis

AMPNet: Asynchronous Model-Parallel Training for Dynamic Neural Networks

1 code implementation ICLR 2018 Alexander L. Gaunt, Matthew A. Johnson, Maik Riechert, Daniel Tarlow, Ryota Tomioka, Dimitrios Vytiniotis, Sam Webster

Through an implementation on multi-core CPUs, we show that AMP training converges to the same accuracy as conventional synchronous training algorithms in a similar number of epochs, but utilizes the available hardware more efficiently even for small minibatch sizes, resulting in significantly shorter overall training times.

Measuring Neural Net Robustness with Constraints

1 code implementation NeurIPS 2016 Osbert Bastani, Yani Ioannou, Leonidas Lampropoulos, Dimitrios Vytiniotis, Aditya Nori, Antonio Criminisi

Despite having high accuracy, neural nets have been shown to be susceptible to adversarial examples, where a small perturbation to an input can cause it to become mislabeled.

Cannot find the paper you are looking for? You can Submit a new open access paper.