no code implementations • 17 Jan 2024 • Kaan Ozkara, Can Karakus, Parameswaran Raman, Mingyi Hong, Shoham Sabach, Branislav Kveton, Volkan Cevher
Since Adam was introduced, several novel adaptive optimizers for deep learning have been proposed.
no code implementations • 10 Nov 2021 • Can Karakus, Rahul Huilgol, Fei Wu, Anirudh Subramanian, Cade Daniel, Derya Cavdar, Teng Xu, Haohan Chen, Arash Rahnama, Luis Quintela
In contrast to existing solutions, the implementation of the SageMaker library is much more generic and flexible, in that it can automatically partition and run pipeline parallelism over arbitrary model architectures with minimal code change, and also offers a general and extensible framework for tensor parallelism, which supports a wider range of use cases, and is modular enough to be easily applied to new training scripts.
1 code implementation • NeurIPS 2019 • Debraj Basu, Deepesh Data, Can Karakus, Suhas Diggavi
Communication bottleneck has been identified as a significant issue in distributed optimization of large-scale learning models.
no code implementations • 6 Jun 2019 • Debraj Basu, Deepesh Data, Can Karakus, Suhas Diggavi
Communication bottleneck has been identified as a significant issue in distributed optimization of large-scale learning models.
no code implementations • 10 May 2019 • Derya Cavdar, Valeriu Codreanu, Can Karakus, John A. Lockman III, Damian Podareanu, Vikram Saletore, Alexander Sergeev, Don D. Smith II, Victor Suthichai, Quy Ta, Srinivas Varadharajan, Lucas A. Wilson, Rengan Xu, Pei Yang
Neural machine translation - using neural networks to translate human language - is an area of active research exploring new neuron types and network topologies with the goal of dramatically improving machine translation performance.
no code implementations • 19 Mar 2019 • Mehrdad Showkatbakhsh, Can Karakus, Suhas Diggavi
Consensus-based optimization consists of a set of computational nodes arranged in a graph, each having a local objective that depends on their local data, where in every step nodes take a linear combination of their neighbors' messages, as well as taking a new gradient step.
no code implementations • 13 Feb 2019 • Mehrdad Showkatbakhsh, Can Karakus, Suhas Diggavi
Data privacy is an important concern in machine learning, and is fundamentally at odds with the task of training useful learning models, which typically require the acquisition of large amounts of private user data.
no code implementations • 14 Mar 2018 • Can Karakus, Yifan Sun, Suhas Diggavi, Wotao Yin
Performance of distributed optimization and learning systems is bottlenecked by "straggler" nodes and slow communication links, which significantly delay computation.
no code implementations • NeurIPS 2017 • Can Karakus, Yifan Sun, Suhas Diggavi, Wotao Yin
Slow running or straggler tasks can significantly reduce computation speed in distributed computation.