no code implementations • 28 Oct 2012 • Robert Nishihara, Iain Murray, Ryan P. Adams
Probabilistic models are conceptually powerful tools for finding structure in data, but their practical effectiveness is often limited by our ability to perform inference in them.
no code implementations • 19 Dec 2013 • Robert Nishihara, Thomas Minka, Daniel Tarlow
Probabilistic models often have parameters that can be translated, scaled, permuted, or otherwise transformed without changing the model.
no code implementations • NeurIPS 2014 • Robert Nishihara, Stefanie Jegelka, Michael. I. Jordan
Submodular functions describe a variety of discrete problems in machine learning, signal processing, and computer vision.
no code implementations • 6 Feb 2015 • Robert Nishihara, Laurent Lessard, Benjamin Recht, Andrew Packard, Michael. I. Jordan
We provide a new proof of the linear convergence of the alternating direction method of multipliers (ADMM) when one of the objective terms is strongly convex.
Optimization and Control Numerical Analysis
1 code implementation • 9 Aug 2015 • Philipp Moritz, Robert Nishihara, Michael. I. Jordan
We propose a new stochastic L-BFGS algorithm and prove a linear convergence rate for strongly convex and smooth functions.
no code implementations • 12 Aug 2015 • Robert Nishihara, David Lopez-Paz, Léon Bottou
This work is naturally framed in the extreme bandit setting, which deals with sequentially choosing which distribution from a collection to sample in order to minimize (maximize) the single best cost (reward).
1 code implementation • 19 Nov 2015 • Philipp Moritz, Robert Nishihara, Ion Stoica, Michael. I. Jordan
We introduce SparkNet, a framework for training deep networks in Spark.
2 code implementations • CVPR 2017 • David Lopez-Paz, Robert Nishihara, Soumith Chintala, Bernhard Schölkopf, Léon Bottou
Our experiments demonstrate the existence of a relation between the direction of causality and the difference between objects and their contexts, and by the same token, the existence of observable signals that reveal the causal dispositions of objects.
2 code implementations • 11 Mar 2017 • Robert Nishihara, Philipp Moritz, Stephanie Wang, Alexey Tumanov, William Paul, Johann Schleier-Smith, Richard Liaw, Mehrdad Niknami, Michael. I. Jordan, Ion Stoica
Machine learning applications are increasingly deployed not only to serve predictions using static models, but also as tightly-integrated components of feedback loops involving dynamic, real-time decision making.
4 code implementations • 16 Dec 2017 • Philipp Moritz, Robert Nishihara, Stephanie Wang, Alexey Tumanov, Richard Liaw, Eric Liang, Melih Elibol, Zongheng Yang, William Paul, Michael. I. Jordan, Ion Stoica
To meet the performance requirements, Ray employs a distributed scheduler and a distributed and fault-tolerant store to manage the system's control state.
3 code implementations • ICML 2018 • Eric Liang, Richard Liaw, Philipp Moritz, Robert Nishihara, Roy Fox, Ken Goldberg, Joseph E. Gonzalez, Michael. I. Jordan, Ion Stoica
Reinforcement learning (RL) algorithms involve the deep nesting of highly irregular computation patterns, each of which typically exhibits opportunities for distributed computation.
4 code implementations • 13 Jul 2018 • Richard Liaw, Eric Liang, Robert Nishihara, Philipp Moritz, Joseph E. Gonzalez, Ion Stoica
We show that this interface meets the requirements for a broad range of hyperparameter search algorithms, allows straightforward scaling of search to large clusters, and simplifies algorithm implementation.
no code implementations • 7 Apr 2019 • Thomas Anthony, Robert Nishihara, Philipp Moritz, Tim Salimans, John Schulman
Monte Carlo Tree Search (MCTS) algorithms perform simulation-based search to improve policies online.
1 code implementation • 13 Feb 2020 • Siyuan Zhuang, Zhuohan Li, Danyang Zhuo, Stephanie Wang, Eric Liang, Robert Nishihara, Philipp Moritz, Ion Stoica
Task-based distributed frameworks (e. g., Ray, Dask, Hydro) have become increasingly popular for distributed applications that contain asynchronous and dynamic workloads, including asynchronous gradient descent, reinforcement learning, and model serving.