no code implementations • 21 Aug 2024 • Tianyi Lin, Chi Jin, Michael. I. Jordan
We provide a unified analysis of two-timescale gradient descent ascent (TTGDA) for solving structured nonconvex minimax optimization problems in the form of $\min_\textbf{x} \max_{\textbf{y} \in Y} f(\textbf{x}, \textbf{y})$, where the objective function $f(\textbf{x}, \textbf{y})$ is nonconvex in $\textbf{x}$ and concave in $\textbf{y}$, and the constraint set $Y \subseteq \mathbb{R}^n$ is convex and bounded.
no code implementations • 6 May 2022 • Tianyi Lin, Michael. I. Jordan
Our method with restarting attains a linear rate for smooth and uniformly monotone VIs and a local superlinear rate for smooth and strongly monotone VIs.
5 code implementations • ICLR 2021 • Anastasios Angelopoulos, Stephen Bates, Jitendra Malik, Michael. I. Jordan
Convolutional image classifiers can achieve high predictive accuracy, but quantifying their uncertainty remains an unresolved challenge, hindering their deployment in consequential settings.
no code implementations • 27 Sep 2020 • Romain Lopez, Inderjit S. Dhillon, Michael. I. Jordan
In POXM, the selected actions for the sIS estimator are the top-p actions of the logging policy, where p is adjusted from the data and is significantly smaller than the size of the action space.
Extreme Multi-Label Classification MUlTI-LABEL-ClASSIFICATION +1
no code implementations • 1 Sep 2020 • Jiri Hron, Karl Krauth, Michael. I. Jordan, Niki Kilbertus
In this work, we focus on the complementary issue of exploration.
no code implementations • 14 Aug 2020 • Yuchen Zhang, Mingsheng Long, Jian-Min Wang, Michael. I. Jordan
Finally, we further extend the localized discrepancies for achieving super transfer and derive generalization bounds that could be even more sample-efficient on source domain.
no code implementations • NeurIPS 2020 • Ximei Wang, Mingsheng Long, Jian-Min Wang, Michael. I. Jordan
In this paper, we delve into the open problem of Calibration in DA, which is extremely challenging due to the coexistence of domain shift and the lack of target labels.
no code implementations • 16 Jul 2020 • Yeshwanth Cherapanamjeri, Efe Aras, Nilesh Tripuraneni, Michael. I. Jordan, Nicolas Flammarion, Peter L. Bartlett
We study the problem of high-dimensional robust linear regression where a learner is given access to $n$ samples from the generative model $Y = \langle X, w^* \rangle + \epsilon$ (with $X \in \mathbb{R}^d$ and $\epsilon$ independent), in which an $\eta$ fraction of the samples have been adversarially corrupted.
no code implementations • 7 Jul 2020 • Daniel Ting, Michael. I. Jordan
Nonlinear dimensionality reduction methods provide a valuable means to visualize and interpret high-dimensional data.
no code implementations • ICML 2020 • Jonathan N. Lee, Aldo Pacchiano, Peter Bartlett, Michael. I. Jordan
Maximum a posteriori (MAP) inference in discrete-valued Markov random fields is a fundamental problem in machine learning that involves identifying the most likely configuration of random variables given a distribution.
no code implementations • 22 Jun 2020 • Tianyi Lin, Zeyu Zheng, Elynn Y. Chen, Marco Cuturi, Michael. I. Jordan
Yet, the behavior of minimum Wasserstein estimators is poorly understood, notably in high-dimensional regimes or under model misspecification.
no code implementations • NeurIPS 2020 • Nilesh Tripuraneni, Michael. I. Jordan, Chi Jin
Formally, we consider $t+1$ tasks parameterized by functions of the form $f_j \circ h$ in a general function class $\mathcal{F} \circ \mathcal{H}$, where each $f_j$ is a task-specific function in $\mathcal{F}$ and $h$ is the shared representation in $\mathcal{H}$.
no code implementations • 18 Jun 2020 • Horia Mania, Michael. I. Jordan, Benjamin Recht
While the identification of nonlinear dynamical systems is a fundamental building block of model-based reinforcement learning and feedback control, its sample complexity is only understood for systems that either have discrete states and actions or for systems that can be identified from data generated by i. i. d.
no code implementations • NeurIPS 2020 • Tianyi Lin, Chenyou Fan, Nhat Ho, Marco Cuturi, Michael. I. Jordan
Projection robust Wasserstein (PRW) distance, or Wasserstein projection pursuit (WPP), is a robust variant of the Wasserstein distance.
no code implementations • 22 May 2020 • Nhat Ho, Koulik Khamaru, Raaz Dwivedi, Martin J. Wainwright, Michael. I. Jordan, Bin Yu
Many statistical estimators are defined as the fixed point of a data-dependent operator, with estimators based on minimizing a cost function being an important special case.
no code implementations • 7 May 2020 • Max Rabinovich, Michael. I. Jordan, Martin J. Wainwright
A line of more recent work in multiple testing has begun to investigate the tradeoffs between the FDR and FNR and to provide lower bounds on the performance of procedures that depend on the model structure.
no code implementations • 19 Apr 2020 • Kirthevasan Kandasamy, Joseph E. Gonzalez, Michael. I. Jordan, Ion Stoica
To that end, we first define three notions of regret for the welfare, the individual utilities of each agent and that of the mechanism.
no code implementations • 15 Apr 2020 • Guilherme França, Michael. I. Jordan, René Vidal
More specifically, we show that a generalization of symplectic integrators to nonconservative and in particular dissipative Hamiltonian systems is able to preserve rates of convergence up to a controlled error.
no code implementations • 15 Apr 2020 • Bin Shi, Weijie J. Su, Michael. I. Jordan
In this paper, we present a general theoretical analysis of the effect of the learning rate in stochastic gradient descent (SGD).
no code implementations • 9 Apr 2020 • Wenlong Mou, Chris Junchi Li, Martin J. Wainwright, Peter L. Bartlett, Michael. I. Jordan
When the matrix $\bar{A}$ is Hurwitz, we prove a central limit theorem (CLT) for the averaged iterates with fixed step size and number of iterations going to infinity.
1 code implementation • 19 Mar 2020 • Anastasios Nikolas Angelopoulos, Reese Pathak, Rohit Varma, Michael. I. Jordan
As we are in the middle of an active outbreak, estimating this measure will necessarily involve correcting for time- and severity- dependent reporting of cases, and time-lags in observed patient outcomes.
no code implementations • 16 Mar 2020 • Koulik Khamaru, Ashwin Pananjady, Feng Ruan, Martin J. Wainwright, Michael. I. Jordan
We address the problem of policy evaluation in discounted Markov decision processes, and provide instance-dependent guarantees on the $\ell_\infty$-error under a generative model.
1 code implementation • 12 Mar 2020 • Esther Rolf, Michael. I. Jordan, Benjamin Recht
Observational data are often accompanied by natural structural indices, such as time stamps or geographic locations, which are meaningful to prediction tasks but are often discarded.
no code implementations • 5 Mar 2020 • Aldo Pacchiano, Heinrich Jiang, Michael. I. Jordan
Mode estimation is a classical problem in statistics with a wide range of applications in machine learning.
no code implementations • 28 Feb 2020 • Michael Muehlebach, Michael. I. Jordan
We analyze the convergence rate of various momentum-based optimization algorithms from a dynamical systems point of view.
1 code implementation • 26 Feb 2020 • Nilesh Tripuraneni, Chi Jin, Michael. I. Jordan
In this paper, we focus on the problem of multi-task linear regression -- in which multiple linear regression models share a common, low-dimensional linear representation.
no code implementations • ICML 2020 • Tianyi Lin, Zhengyuan Zhou, Panayotis Mertikopoulos, Michael. I. Jordan
In this paper, we consider multi-agent learning via online gradient descent in a class of games called $\lambda$-cocoercive games, a fairly broad class of games that admits many Nash equilibria and that properly includes unconstrained strongly monotone games.
no code implementations • ICML 2020 • Eric Mazumdar, Aldo Pacchiano, Yi-An Ma, Peter L. Bartlett, Michael. I. Jordan
The resulting approximate Thompson sampling algorithm has logarithmic regret and its computational complexity does not scale with the time horizon of the algorithm.
1 code implementation • NeurIPS 2020 • Serena Wang, Wenshuo Guo, Harikrishna Narasimhan, Andrew Cotter, Maya Gupta, Michael. I. Jordan
Second, we introduce two new approaches using robust optimization that, unlike the naive approach of only relying on $\hat{G}$, are guaranteed to satisfy fairness criteria on the true protected groups G while minimizing a training objective.
2 code implementations • NeurIPS 2020 • Romain Lopez, Pierre Boyeau, Nir Yosef, Michael. I. Jordan, Jeffrey Regier
To make decisions based on a model fit with auto-encoding variational Bayes (AEVB), practitioners often let the variational distribution serve as a surrogate for the posterior distribution.
1 code implementation • 13 Feb 2020 • Samuel Horváth, Lihua Lei, Peter Richtárik, Michael. I. Jordan
Adaptivity is an important yet under-studied property in modern optimization theory.
no code implementations • NeurIPS 2020 • Tianyi Lin, Nhat Ho, Xi Chen, Marco Cuturi, Michael. I. Jordan
We study the fixed-support Wasserstein barycenter problem (FS-WBP), which consists in computing the Wasserstein barycenter of $m$ discrete probability measures supported on a finite metric space of size $n$.
no code implementations • ICML 2020 • Michael Muehlebach, Michael. I. Jordan
This article derives lower bounds on the convergence rate of continuous-time gradient-based optimization algorithms.
Optimization and Control Systems and Control Systems and Control
no code implementations • 5 Feb 2020 • Tianyi Lin, Chi Jin, Michael. I. Jordan
This paper presents the first algorithm with $\tilde{O}(\sqrt{\kappa_{\mathbf x}\kappa_{\mathbf y}})$ gradient complexity, matching the lower bound up to logarithmic factors.
no code implementations • ICLR 2020 • Melih Elibol, Lihua Lei, Michael. I. Jordan
Variance reduction methods such as SVRG and SpiderBoost use a mixture of large and small batch gradients to reduce the variance of stochastic gradients.
no code implementations • 11 Dec 2019 • Wenlong Mou, Nhat Ho, Martin J. Wainwright, Peter L. Bartlett, Michael. I. Jordan
We study the problem of sampling from the power posterior distribution in Bayesian Gaussian mixture models, a robust version of the classical posterior.
1 code implementation • NeurIPS 2019 • Ximei Wang, Ying Jin, Mingsheng Long, Jian-Min Wang, Michael. I. Jordan
Deep neural networks (DNNs) excel at learning representations when trained on large-scale datasets.
no code implementations • 11 Oct 2019 • Tijana Zrnic, Daniel L. Jiang, Aaditya Ramdas, Michael. I. Jordan
One important partition of algorithms for controlling the false discovery rate (FDR) in multiple testing is into offline and online algorithms.
1 code implementation • 30 Sep 2019 • Tianyi Lin, Nhat Ho, Marco Cuturi, Michael. I. Jordan
This provides a first \textit{near-linear time} complexity bound guarantee for approximating the MOT problem and matches the best known complexity bound for the Sinkhorn algorithm in the classical OT setting when $m = 2$.
no code implementations • 26 Sep 2019 • Hong Liu, Mingsheng Long, Jian-Min Wang, Michael. I. Jordan
3) The feasibility of transferability is related to the similarity of both input and label.
no code implementations • 28 Aug 2019 • Wenlong Mou, Yi-An Ma, Martin J. Wainwright, Peter L. Bartlett, Michael. I. Jordan
We propose a Markov chain Monte Carlo (MCMC) algorithm based on third-order Langevin dynamics for sampling from distributions with log-concave and smooth densities.
no code implementations • ICLR 2020 • Kaichao You, Mingsheng Long, Jian-Min Wang, Michael. I. Jordan
Despite the popularity of these common beliefs, experiments suggest that they are insufficient in explaining the general effectiveness of lrDecay in training modern neural networks that are deep, wide, and nonconvex.
1 code implementation • 28 Jul 2019 • Ryan Giordano, Michael. I. Jordan, Tamara Broderick
The first-order approximation is known as the "infinitesimal jackknife" in the statistics literature and has been the subject of recent interest in machine learning for approximate CV.
no code implementations • 27 Jul 2019 • Kush Bhatia, Yi-An Ma, Anca D. Dragan, Peter L. Bartlett, Michael. I. Jordan
We study the problem of robustly estimating the posterior distribution for the setting where observed data can be contaminated with potentially adversarial outliers.
2 code implementations • 11 Jul 2019 • Chi Jin, Zhuoran Yang, Zhaoran Wang, Michael. I. Jordan
Modern Reinforcement Learning (RL) is commonly applied to practical problems with an enormous number of states, where function approximation must be deployed to approximate either the value function or the policy.
no code implementations • 9 Jul 2019 • Nhat Ho, Chiao-Yu Yang, Michael. I. Jordan
We provide a theoretical treatment of over-specified Gaussian mixtures of experts with covariate-free gating networks.
no code implementations • 8 Jul 2019 • Eric Mazumdar, Lillian J. Ratliff, Michael. I. Jordan, S. Shankar Sastry
In such games the state and action spaces are continuous and global Nash equilibria can be found be solving coupled Ricatti equations.
no code implementations • ICML 2020 • Xiang Cheng, Dong Yin, Peter L. Bartlett, Michael. I. Jordan
We prove quantitative convergence rates at which discrete Langevin-like processes converge to the invariant distribution of a related stochastic differential equation.
no code implementations • 2 Jul 2019 • Jonathan N. Lee, Aldo Pacchiano, Michael. I. Jordan
Maximum a posteriori (MAP) inference is a fundamental computational paradigm for statistical inference.
no code implementations • 12 Jun 2019 • Lydia T. Liu, Horia Mania, Michael. I. Jordan
Stable matching, a classical model for two-sided markets, has long been studied with little consideration for how each side's preferences are learned.
1 code implementation • ICML 2020 • Aldo Pacchiano, Jack Parker-Holder, Yunhao Tang, Anna Choromanska, Krzysztof Choromanski, Michael. I. Jordan
We introduce a new approach for comparing reinforcement learning policies, using Wasserstein distances (WDs) in a newly defined latent behavioral space.
no code implementations • 8 Jun 2019 • Puyudi Yang, Jianbo Chen, Cho-Jui Hsieh, Jane-Ling Wang, Michael. I. Jordan
Furthermore, we extend our method to include multi-layer feature attributions in order to tackle the attacks with mixed confidence levels.
no code implementations • ICML 2020 • Tianyi Lin, Chi Jin, Michael. I. Jordan
We consider nonconvex-concave minimax problems, $\min_{\mathbf{x}} \max_{\mathbf{y} \in \mathcal{Y}} f(\mathbf{x}, \mathbf{y})$, where $f$ is nonconvex in $\mathbf{x}$ but concave in $\mathbf{y}$ and $\mathcal{Y}$ is a convex and bounded set.
no code implementations • 2 Jun 2019 • Jelena Diakonikolas, Michael. I. Jordan
We take a Hamiltonian-based perspective to generalize Nesterov's accelerated gradient descent and Polyak's heavy ball method to a broad class of momentum methods in the setting of (possibly) constrained minimization in Euclidean and non-Euclidean normed vector spaces.
no code implementations • 1 Jun 2019 • Tianyi Lin, Nhat Ho, Michael. I. Jordan
We prove that APDAMD achieves the complexity bound of $\widetilde{O}(n^2\sqrt{\delta}\varepsilon^{-1})$ in which $\delta>0$ stands for the regularity of $\phi$.
no code implementations • 30 May 2019 • Niladri S. Chatterji, Jelena Diakonikolas, Michael. I. Jordan, Peter L. Bartlett
Langevin Monte Carlo (LMC) is an iterative algorithm used to generate samples from a distribution that is known only up to a normalizing constant.
no code implementations • 23 May 2019 • Wenshuo Guo, Nhat Ho, Michael. I. Jordan
First, we introduce the \emph{accelerated primal-dual randomized coordinate descent} (APDRCD) algorithm for computing the OT distance.
no code implementations • 17 May 2019 • Michael Muehlebach, Michael. I. Jordan
We present a dynamical system framework for understanding Nesterov's accelerated gradient method.
2 code implementations • 6 May 2019 • Romain Lopez, Achille Nazaret, Maxime Langevin, Jules Samaran, Jeffrey Regier, Michael. I. Jordan, Nir Yosef
Building upon domain adaptation work, we propose gimVI, a deep generative model for the integration of spatial transcriptomic data and scRNA-seq data that can be used to impute missing genes.
no code implementations • ICLR 2019 • Nhat Ho, Tan Nguyen, Ankit B. Patel, Anima Anandkumar, Michael. I. Jordan, Richard G. Baraniuk
The conjugate prior yields a new regularizer for learning based on the paths rendered in the generative model for training CNNs–the Rendering Path Normalization (RPN).
no code implementations • 16 Apr 2019 • Nhat Ho, Tianyi Lin, Michael. I. Jordan
We also conduct experiments on real datasets and the numerical results demonstrate the effectiveness of our algorithms.
6 code implementations • 11 Apr 2019 • Yuchen Zhang, Tianle Liu, Mingsheng Long, Michael. I. Jordan
We introduce Margin Disparity Discrepancy, a novel measurement with rigorous generalization bounds, tailored to the distribution comparison with the asymmetric margin loss, and to the minimax optimization for easier training.
no code implementations • 9 Apr 2019 • Lihua Lei, Michael. I. Jordan
Stochastic-gradient-based optimization has been a core enabling methodology in applications to large-scale problems in machine learning and related areas.
3 code implementations • 3 Apr 2019 • Jianbo Chen, Michael. I. Jordan, Martin J. Wainwright
We develop HopSkipJumpAttack, a family of algorithms based on a novel estimate of the gradient direction using binary information at the decision boundary.
no code implementations • 29 Mar 2019 • Alexander Ratner, Dan Alistarh, Gustavo Alonso, David G. Andersen, Peter Bailis, Sarah Bird, Nicholas Carlini, Bryan Catanzaro, Jennifer Chayes, Eric Chung, Bill Dally, Jeff Dean, Inderjit S. Dhillon, Alexandros Dimakis, Pradeep Dubey, Charles Elkan, Grigori Fursin, Gregory R. Ganger, Lise Getoor, Phillip B. Gibbons, Garth A. Gibson, Joseph E. Gonzalez, Justin Gottschlich, Song Han, Kim Hazelwood, Furong Huang, Martin Jaggi, Kevin Jamieson, Michael. I. Jordan, Gauri Joshi, Rania Khalaf, Jason Knight, Jakub Konečný, Tim Kraska, Arun Kumar, Anastasios Kyrillidis, Aparna Lakshmiratan, Jing Li, Samuel Madden, H. Brendan McMahan, Erik Meijer, Ioannis Mitliagkas, Rajat Monga, Derek Murray, Kunle Olukotun, Dimitris Papailiopoulos, Gennady Pekhimenko, Theodoros Rekatsinas, Afshin Rostamizadeh, Christopher Ré, Christopher De Sa, Hanie Sedghi, Siddhartha Sen, Virginia Smith, Alex Smola, Dawn Song, Evan Sparks, Ion Stoica, Vivienne Sze, Madeleine Udell, Joaquin Vanschoren, Shivaram Venkataraman, Rashmi Vinayak, Markus Weimer, Andrew Gordon Wilson, Eric Xing, Matei Zaharia, Ce Zhang, Ameet Talwalkar
Machine learning (ML) techniques are enjoying rapidly increasing adoption.
no code implementations • 13 Feb 2019 • Chi Jin, Praneeth Netrapalli, Rong Ge, Sham M. Kakade, Michael. I. Jordan
More recent theory has shown that GD and SGD can avoid saddle points, but the dependence on dimension in these analyses is polynomial.
no code implementations • NeurIPS 2019 • Bin Shi, Simon S. Du, Weijie J. Su, Michael. I. Jordan
We study first-order optimization methods obtained by discretizing ordinary differential equations (ODEs) corresponding to Nesterov's accelerated gradient methods (NAGs) and Polyak's heavy-ball method.
no code implementations • 11 Feb 2019 • Chi Jin, Praneeth Netrapalli, Rong Ge, Sham M. Kakade, Michael. I. Jordan
In this note, we derive concentration inequalities for random vectors with subGaussian norm (a generalization of both subGaussian random vectors and norm bounded random vectors), which are tight up to logarithmic factors.
no code implementations • 11 Feb 2019 • Jianbo Chen, Michael. I. Jordan
We study the problem of interpreting trained classification models in the setting of linguistic data sets.
no code implementations • 7 Feb 2019 • Romain Lopez, Chenchen Li, Xiang Yan, Junwu Xiong, Michael. I. Jordan, Yuan Qi, Le Song
We address a practical problem ubiquitous in modern marketing campaigns, in which a central agent tries to learn a policy for allocating strategic financial incentives to customers and observes only bandit feedback.
no code implementations • 4 Feb 2019 • Yi-An Ma, Niladri Chatterji, Xiang Cheng, Nicolas Flammarion, Peter Bartlett, Michael. I. Jordan
We formulate gradient-based Markov chain Monte Carlo (MCMC) sampling as optimization on the space of probability measures, with Kullback-Leibler (KL) divergence as the objective functional.
no code implementations • 3 Feb 2019 • Xiang Cheng, Peter L. Bartlett, Michael. I. Jordan
In this paper, we quantitative convergence in $W_2$ for a family of Langevin-like stochastic processes that includes stochastic gradient descent and related gradient-based algorithms.
1 code implementation • ICML 2020 • Chi Jin, Praneeth Netrapalli, Michael. I. Jordan
Minimax optimization has found extensive applications in modern machine learning, in settings such as generative adversarial networks (GANs), adversarial training and multi-agent reinforcement learning.
BIG-bench Machine Learning Multi-agent Reinforcement Learning +1
no code implementations • 1 Feb 2019 • Raaz Dwivedi, Nhat Ho, Koulik Khamaru, Martin J. Wainwright, Michael. I. Jordan, Bin Yu
We study a class of weakly identifiable location-scale mixture models for which the maximum likelihood estimates based on $n$ i. i. d.
9 code implementations • 24 Jan 2019 • Hongyang Zhang, Yaodong Yu, Jiantao Jiao, Eric P. Xing, Laurent El Ghaoui, Michael. I. Jordan
We identify a trade-off between robustness and accuracy that serves as a guiding principle in the design of defenses against adversarial examples.
Ranked #3 on Adversarial Attack on CIFAR-10
no code implementations • 19 Jan 2019 • Tianyi Lin, Nhat Ho, Michael. I. Jordan
We show that a greedy variant of the classical Sinkhorn algorithm, known as the \emph{Greenkhorn algorithm}, can be improved to $\widetilde{\mathcal{O}}(n^2\varepsilon^{-2})$, improving on the best known complexity bound of $\widetilde{\mathcal{O}}(n^2\varepsilon^{-3})$.
Data Structures and Algorithms
no code implementations • 3 Jan 2019 • Eric V. Mazumdar, Michael. I. Jordan, S. Shankar Sastry
We propose local symplectic surgery, a two-timescale procedure for finding local Nash equilibria in two-player zero-sum games.
2 code implementations • 12 Dec 2018 • Tijana Zrnic, Aaditya Ramdas, Michael. I. Jordan
We consider the problem of asynchronous online testing, aimed at providing control of the false discovery rate (FDR) during a continual stream of data collection and testing, where each test may be a sequential test that can start and stop at arbitrary times.
no code implementations • NeurIPS 2018 • Raaz Dwivedi, Nhật Hồ, Koulik Khamaru, Martin J. Wainwright, Michael. I. Jordan
We provide two classes of theoretical guarantees: first, we characterize the bias introduced due to the misspecification; and second, we prove that population EM converges at a geometric rate to the model projection under a suitable initialization condition.
no code implementations • NeurIPS 2018 • Shichen Liu, Mingsheng Long, Jian-Min Wang, Michael. I. Jordan
A technical challenge of deep learning is recognizing target classes without seen data.
no code implementations • NeurIPS 2018 • Kush Bhatia, Aldo Pacchiano, Nicolas Flammarion, Peter L. Bartlett, Michael. I. Jordan
In this paper, we study the problems of principle Generalized Eigenvector computation and Canonical Correlation Analysis in the stochastic setting.
no code implementations • 20 Nov 2018 • Yi-An Ma, Yuansi Chen, Chi Jin, Nicolas Flammarion, Michael. I. Jordan
Optimization algorithms and Monte Carlo sampling algorithms have provided the computational foundations for the rapid growth in applications of statistical machine learning in recent years.
no code implementations • 20 Nov 2018 • Kush Bhatia, Aldo Pacchiano, Nicolas Flammarion, Peter L. Bartlett, Michael. I. Jordan
In this paper, we study the problems of principal Generalized Eigenvector computation and Canonical Correlation Analysis in the stochastic setting.
no code implementations • 1 Nov 2018 • Tan Nguyen, Nhat Ho, Ankit Patel, Anima Anandkumar, Michael. I. Jordan, Richard G. Baraniuk
This conjugate prior yields a new regularizer based on paths rendered in the generative model for training CNNs-the Rendering Path Normalization (RPN).
no code implementations • 29 Oct 2018 • Nhat Ho, Viet Huynh, Dinh Phung, Michael. I. Jordan
We propose a novel probabilistic approach to multilevel clustering problems based on composite transportation distance, which is a variant of transportation distance where the underlying metric is Kullback-Leibler divergence.
no code implementations • 21 Oct 2018 • Bin Shi, Simon S. Du, Michael. I. Jordan, Weijie J. Su
We also show that these ODEs are more accurate surrogates for the underlying algorithms; in particular, they not only distinguish between NAG-SC and Polyak's heavy-ball method, but they allow the identification of a term that we refer to as "gradient correction" that is present in NAG-SC but not in the heavy-ball method and is responsible for the qualitative difference in convergence of the two methods.
4 code implementations • 15 Oct 2018 • Runjing Liu, Ryan Giordano, Michael. I. Jordan, Tamara Broderick
Bayesian models based on the Dirichlet process and other stick-breaking priors have been proposed as core ingredients for clustering, topic modeling, and other unsupervised learning tasks.
Methodology
1 code implementation • 10 Oct 2018 • Runjing Liu, Jeffrey Regier, Nilesh Tripuraneni, Michael. I. Jordan, Jon McAuliffe
We wish to compute the gradient of an expectation over a finite or countably infinite sample space having $K \leq \infty$ categories.
no code implementations • 1 Oct 2018 • Raaz Dwivedi, Nhat Ho, Koulik Khamaru, Michael. I. Jordan, Martin J. Wainwright, Bin Yu
A line of recent work has analyzed the behavior of the Expectation-Maximization (EM) algorithm in the well-specified setting, in which the population likelihood is locally strongly concave around its maximizing argument.
1 code implementation • 16 Sep 2018 • Maxime Langevin, Edouard Mehlman, Jeffrey Regier, Romain Lopez, Michael. I. Jordan, Nir Yosef
Class labels are often imperfectly observed, due to mistakes and to genuine ambiguity among classes.
1 code implementation • ICLR 2019 • Jianbo Chen, Le Song, Martin J. Wainwright, Michael. I. Jordan
We study instancewise feature importance scoring as a method for model interpretation.
1 code implementation • NeurIPS 2018 • Chi Jin, Zeyuan Allen-Zhu, Sebastien Bubeck, Michael. I. Jordan
We prove that, in an episodic MDP setting, Q-learning with UCB exploration achieves regret $\tilde{O}(\sqrt{H^3 SAT})$, where $S$ and $A$ are the numbers of states and actions, $H$ is the number of steps per episode, and $T$ is the total number of steps.
no code implementations • 25 Jun 2018 • Ahmed El Alaoui, Florent Krzakala, Michael. I. Jordan
We study the fundamental limits of detecting the presence of an additive rank-one perturbation, or spike, to a Wigner matrix.
no code implementations • 1 Jun 2018 • Tianyi Lin, Chenyou Fan, Mengdi Wang, Michael. I. Jordan
Convex composition optimization is an emerging topic that covers a wide range of applications arising from stochastic optimal control, reinforcement learning and multi-stage stochastic programming.
3 code implementations • 1 Jun 2018 • Ryan Giordano, Will Stephenson, Runjing Liu, Michael. I. Jordan, Tamara Broderick
This linear approximation is sometimes known as the "infinitesimal jackknife" in the statistics literature, where it is mostly used to as a theoretical tool to prove asymptotic results.
Methodology
no code implementations • 31 May 2018 • Puyudi Yang, Jianbo Chen, Cho-Jui Hsieh, Jane-Ling Wang, Michael. I. Jordan
We present a probabilistic framework for studying adversarial attacks on discrete data.
no code implementations • NeurIPS 2018 • Romain Lopez, Jeffrey Regier, Michael. I. Jordan, Nir Yosef
We show how to apply this method to a range of problems, including the problems of learning invariant representations and the learning of interpretable representations.
no code implementations • 4 May 2018 • Xiang Cheng, Niladri S. Chatterji, Yasin Abbasi-Yadkori, Peter L. Bartlett, Michael. I. Jordan
We study the problem of sampling from a distribution $p^*(x) \propto \exp\left(-U(x)\right)$, where the function $U$ is $L$-smooth everywhere and $m$-strongly convex outside a ball of radius $R$, but potentially nonconvex inside this ball.
no code implementations • NeurIPS 2018 • Chi Jin, Lydia T. Liu, Rong Ge, Michael. I. Jordan
Our objective is to find the $\epsilon$-approximate local minima of the underlying function $F$ while avoiding the shallow local minima---arising because of the tolerance $\nu$---which exist only in $f$.
no code implementations • 6 Mar 2018 • Daniel Ting, Michael. I. Jordan
There is limited basis for using existing NLDR theory for deriving new algorithms.
no code implementations • 28 Feb 2018 • Vladimir Feinberg, Alvin Wan, Ion Stoica, Michael. I. Jordan, Joseph E. Gonzalez, Sergey Levine
By enabling wider use of learned dynamics models within a model-free reinforcement learning algorithm, we improve value estimation, which, in turn, reduces the sample complexity of learning.
no code implementations • 26 Feb 2018 • Nilesh Tripuraneni, Nicolas Flammarion, Francis Bach, Michael. I. Jordan
We consider the minimization of a function defined on a Riemannian manifold $\mathcal{M}$ accessible only through unbiased estimates of its gradients.
no code implementations • 22 Feb 2018 • Max Simchowitz, Horia Mania, Stephen Tu, Michael. I. Jordan, Benjamin Recht
We prove that the ordinary least-squares (OLS) estimator attains nearly minimax optimal performance for the identification of linear dynamical systems from a single observed trajectory.
3 code implementations • ICML 2018 • Jianbo Chen, Le Song, Martin J. Wainwright, Michael. I. Jordan
We introduce instancewise feature selection as a methodology for model interpretation.
no code implementations • 20 Feb 2018 • Ahmed El Alaoui, Michael. I. Jordan
This region is shaped by the prior in a non-trivial way.
no code implementations • ICML 2018 • Niladri S. Chatterji, Nicolas Flammarion, Yi-An Ma, Peter L. Bartlett, Michael. I. Jordan
We provide convergence guarantees in Wasserstein distance for a variety of variance-reduction methods: SAGA Langevin diffusion, SVRG Langevin diffusion and control-variate underdamped Langevin diffusion.
3 code implementations • ICML 2018 • Eric Liang, Richard Liaw, Philipp Moritz, Robert Nishihara, Roy Fox, Ken Goldberg, Joseph E. Gonzalez, Michael. I. Jordan, Ion Stoica
Reinforcement learning (RL) algorithms involve the deep nesting of highly irregular computation patterns, each of which typically exhibits opportunities for distributed computation.
4 code implementations • 16 Dec 2017 • Philipp Moritz, Robert Nishihara, Stephanie Wang, Alexey Tumanov, Richard Liaw, Eric Liang, Melih Elibol, Zongheng Yang, William Paul, Michael. I. Jordan, Ion Stoica
To meet the performance requirements, Ray employs a distributed scheduler and a distributed and fault-tolerant store to manage the system's control state.
no code implementations • 28 Nov 2017 • Chi Jin, Praneeth Netrapalli, Michael. I. Jordan
Nesterov's accelerated gradient descent (AGD), an instance of the general family of "momentum methods", provably achieves faster convergence rate than gradient descent (GD) in the convex setting.
no code implementations • NeurIPS 2018 • Nilesh Tripuraneni, Mitchell Stern, Chi Jin, Jeffrey Regier, Michael. I. Jordan
This paper proposes a stochastic variant of a classic algorithm---the cubic-regularized Newton method [Nesterov and Polyak 2006].
no code implementations • 20 Oct 2017 • Jason D. Lee, Ioannis Panageas, Georgios Piliouras, Max Simchowitz, Michael. I. Jordan, Benjamin Recht
We establish that first-order methods avoid saddle points for almost all initializations.
1 code implementation • NeurIPS 2017 • Aaditya Ramdas, Fanny Yang, Martin J. Wainwright, Michael. I. Jordan
In the online multiple testing problem, p-values corresponding to different null hypotheses are observed one by one, and the decision of whether or not to reject the current hypothesis must be made immediately, after which the next p-value is observed.
1 code implementation • 29 Sep 2017 • Aaditya Ramdas, Jianbo Chen, Martin J. Wainwright, Michael. I. Jordan
We propose a linear-time, single-pass, top-down algorithm for multiple testing on directed acyclic graphs (DAGs), where nodes represent hypotheses and edges specify a partial ordering in which hypotheses must be tested.
4 code implementations • 8 Sep 2017 • Ryan Giordano, Tamara Broderick, Michael. I. Jordan
The estimates for MFVB posterior covariances rely on a result from the classical Bayesian robustness literature relating derivatives of posterior expectations to posterior covariances and include the Laplace approximation as a special case.
Methodology
no code implementations • CVPR 2018 • Zhangjie Cao, Mingsheng Long, Jian-Min Wang, Michael. I. Jordan
Existing domain adversarial networks assume fully shared label space across domains.
no code implementations • 12 Jul 2017 • Xiang Cheng, Niladri S. Chatterji, Peter L. Bartlett, Michael. I. Jordan
We study the underdamped Langevin diffusion when the log of the target distribution is smooth and strongly concave.
1 code implementation • NeurIPS 2017 • Jianbo Chen, Mitchell Stern, Martin J. Wainwright, Michael. I. Jordan
We propose a method for feature selection that employs kernel-based measures of independence to find a subset of covariates that is maximally predictive of the response.
no code implementations • NeurIPS 2017 • Jeffrey Regier, Michael. I. Jordan, Jon McAuliffe
We introduce TrustVI, a fast second-order algorithm for black-box variational inference based on trust-region optimization and the reparameterization trick.
no code implementations • NeurIPS 2017 • Simon S. Du, Chi Jin, Jason D. Lee, Michael. I. Jordan, Barnabas Poczos, Aarti Singh
Although gradient descent (GD) almost always escapes saddle points asymptotically [Lee et al., 2016], this paper shows that even with fairly natural random initialization schemes and non-pathological functions, GD can be significantly slowed down by saddle points, taking exponential time to escape.
5 code implementations • NeurIPS 2018 • Mingsheng Long, Zhangjie Cao, Jian-Min Wang, Michael. I. Jordan
Adversarial learning has been embedded into deep networks to learn disentangled and transferable representations for domain adaptation.
Ranked #7 on Domain Adaptation on USPS-to-MNIST
no code implementations • 18 Mar 2017 • Aaditya Ramdas, Rina Foygel Barber, Martin J. Wainwright, Michael. I. Jordan
There is a significant literature on methods for incorporating knowledge into multiple testing procedures so as to improve their power and precision.
2 code implementations • 11 Mar 2017 • Robert Nishihara, Philipp Moritz, Stephanie Wang, Alexey Tumanov, William Paul, Johann Schleier-Smith, Richard Liaw, Mehrdad Niknami, Michael. I. Jordan, Ion Stoica
Machine learning applications are increasingly deployed not only to serve predictions using static models, but also as tightly-integrated components of feedback loops involving dynamic, real-time decision making.
no code implementations • ICML 2017 • Chi Jin, Rong Ge, Praneeth Netrapalli, Sham M. Kakade, Michael. I. Jordan
This paper shows that a perturbed form of gradient descent converges to a second-order stationary point in a number iterations which depends only poly-logarithmically on dimension (i. e., it is almost "dimension-free").
2 code implementations • 7 Nov 2016 • Virginia Smith, Simone Forte, Chenxin Ma, Martin Takac, Michael. I. Jordan, Martin Jaggi
The scale of modern datasets necessitates the development of efficient distributed optimization methods for machine learning.
no code implementations • 12 Sep 2016 • Lihua Lei, Michael. I. Jordan
We develop and analyze a procedure for gradient-based optimization that we refer to as stochastically controlled stochastic gradient (SCSG).
1 code implementation • NeurIPS 2016 • Xinghao Pan, Maximilian Lam, Stephen Tu, Dimitris Papailiopoulos, Ce Zhang, Michael. I. Jordan, Kannan Ramchandran, Chris Re, Benjamin Recht
We present CYCLADES, a general framework for parallelizing stochastic optimization algorithms in a shared memory setting.
no code implementations • 25 May 2016 • Michael. I. Jordan, Jason D. Lee, Yun Yang
CSL provides a communication-efficient surrogate to the global likelihood that can be used for low-dimensional estimation, high-dimensional regularized estimation and Bayesian inference.
1 code implementation • ICML 2017 • Mingsheng Long, Han Zhu, Jian-Min Wang, Michael. I. Jordan
Deep networks have been successfully applied to learn transferable features for adapting models from a source domain to a different target domain.
Ranked #2 on Domain Adaptation on HMDBfull-to-UCF
Multi-Source Unsupervised Domain Adaptation Transfer Learning
no code implementations • 6 May 2016 • Maxim Rabinovich, Aaditya Ramdas, Michael. I. Jordan, Martin J. Wainwright
These results show that it is possible for empirical expectations of functions to concentrate long before the underlying chain has mixed in the classical sense, and we show that the concentration rates we achieve are optimal up to constants.
no code implementations • 25 Mar 2016 • Horia Mania, Aaditya Ramdas, Martin J. Wainwright, Michael. I. Jordan, Benjamin Recht
This paper studies the use of reproducing kernel Hilbert space methods for learning from permutation-valued features.
no code implementations • 14 Mar 2016 • Andre Wibisono, Ashia C. Wilson, Michael. I. Jordan
We show that there is a Lagrangian functional that we call the \emph{Bregman Lagrangian} which generates a large class of accelerated methods in continuous time, including (but not limited to) accelerated gradient descent, its non-Euclidean extension, and accelerated higher-order gradient methods.
no code implementations • 2 Mar 2016 • Ahmed El Alaoui, Xiang Cheng, Aaditya Ramdas, Martin J. Wainwright, Michael. I. Jordan
Together, these properties show that $p = d+1$ is an optimal choice, yielding a function estimate $\hat{f}$ that is both smooth and non-degenerate, while remaining maximally sensitive to $P$.
no code implementations • 16 Feb 2016 • Jason D. Lee, Max Simchowitz, Michael. I. Jordan, Benjamin Recht
We show that gradient descent converges to a local minimizer, almost surely with random initialization.
2 code implementations • NeurIPS 2016 • Mingsheng Long, Han Zhu, Jian-Min Wang, Michael. I. Jordan
In this paper, we propose a new approach to domain adaptation in deep networks that can jointly learn adaptive classifiers and transferable features from labeled data in the source domain and unlabeled data in the target domain.
no code implementations • 10 Feb 2016 • Qiang Liu, Jason D. Lee, Michael. I. Jordan
We derive a new discrepancy statistic for measuring differences between two probability distributions based on combining Stein's identity with the reproducing kernel Hilbert space theory.
2 code implementations • 13 Dec 2015 • Virginia Smith, Simone Forte, Michael. I. Jordan, Martin Jaggi
Despite the importance of sparsity in many large-scale applications, there are few methods for distributed optimization of sparsity-inducing objectives.
1 code implementation • 13 Dec 2015 • Chenxin Ma, Jakub Konečný, Martin Jaggi, Virginia Smith, Michael. I. Jordan, Peter Richtárik, Martin Takáč
To this end, we present a framework for distributed optimization that both allows the flexibility of arbitrary solvers to be used on each (single) machine locally, and yet maintains competitive performance against other state-of-the-art special-purpose distributed methods.
no code implementations • 25 Nov 2015 • Yuchen Zhang, Jason D. Lee, Martin J. Wainwright, Michael. I. Jordan
For loss functions that are $L$-Lipschitz continuous, we present algorithms to learn halfspaces and multi-layer neural networks that achieve arbitrarily small excess risk $\epsilon>0$.
1 code implementation • 19 Nov 2015 • Philipp Moritz, Robert Nishihara, Ion Stoica, Michael. I. Jordan
We introduce SparkNet, a framework for training deep networks in Spark.
no code implementations • 13 Oct 2015 • Yuchen Zhang, Jason D. Lee, Michael. I. Jordan
The sample complexity and the time complexity of the presented method are polynomial in the input dimension and in $(1/\epsilon,\log(1/\delta), F(k, L))$, where $F(k, L)$ is a function depending on $(k, L)$ and on the activation function, independent of the number of neurons.
1 code implementation • 9 Aug 2015 • Philipp Moritz, Robert Nishihara, Michael. I. Jordan
We propose a new stochastic L-BFGS algorithm and prove a linear convergence rate for strongly convex and smooth functions.
no code implementations • 24 Jul 2015 • Horia Mania, Xinghao Pan, Dimitris Papailiopoulos, Benjamin Recht, Kannan Ramchandran, Michael. I. Jordan
We demonstrate experimentally on a 16-core machine that the sparse and parallel version of SVRG is in some cases more than four orders of magnitude faster than the standard SVRG algorithm.
no code implementations • NeurIPS 2015 • Xinghao Pan, Dimitris Papailiopoulos, Samet Oymak, Benjamin Recht, Kannan Ramchandran, Michael. I. Jordan
We present C4 and ClusterWild!, two algorithms for parallel correlation clustering that run in a polylogarithmic number of rounds and achieve nearly linear speedups, provably.
no code implementations • 24 Jun 2015 • Yuchen Zhang, Michael. I. Jordan
Splash consists of a programming interface and an execution engine.
no code implementations • NeurIPS 2015 • Jacob Andreas, Maxim Rabinovich, Dan Klein, Michael. I. Jordan
Calculation of the log-normalizer is a major computational obstacle in applications of log-linear models with large output spaces.
no code implementations • NeurIPS 2015 • Maxim Rabinovich, Elaine Angelino, Michael. I. Jordan
Practitioners of Bayesian statistics have long depended on Markov chain Monte Carlo (MCMC) to obtain samples from intractable posterior distributions.
no code implementations • 29 May 2015 • Yun Yang, Martin J. Wainwright, Michael. I. Jordan
We study the computational complexity of Markov chain Monte Carlo (MCMC) methods for high-dimensional Bayesian linear regression under sparsity constraints.
no code implementations • 11 Mar 2015 • Yuchen Zhang, Martin J. Wainwright, Michael. I. Jordan
In this paper, we show that the slow rate is intrinsic to a broad class of M-estimators.
22 code implementations • 19 Feb 2015 • John Schulman, Sergey Levine, Philipp Moritz, Michael. I. Jordan, Pieter Abbeel
We describe an iterative procedure for optimizing policies, with guaranteed monotonic improvement.
1 code implementation • 12 Feb 2015 • Chenxin Ma, Virginia Smith, Martin Jaggi, Michael. I. Jordan, Peter Richtárik, Martin Takáč
Distributed optimization methods for large-scale machine learning suffer from a communication bottleneck.
6 code implementations • 10 Feb 2015 • Mingsheng Long, Yue Cao, Jian-Min Wang, Michael. I. Jordan
Recent studies reveal that a deep neural network can learn transferable features which generalize well to novel tasks for domain adaptation.
Ranked #3 on Domain Adaptation on Synth Digits-to-SVHN
no code implementations • 6 Feb 2015 • Robert Nishihara, Laurent Lessard, Benjamin Recht, Andrew Packard, Michael. I. Jordan
We provide a new proof of the linear convergence of the alternating direction method of multipliers (ADMM) when one of the objective terms is strongly convex.
Optimization and Control Numerical Analysis
no code implementations • 5 Feb 2015 • Yuchen Zhang, Martin J. Wainwright, Michael. I. Jordan
We study the following generalized matrix rank estimation problem: given an $n \times n$ matrix and a constant $c \geq 0$, estimate the number of eigenvalues that are greater than $c$.
no code implementations • 31 Jan 2015 • Evan R. Sparks, Ameet Talwalkar, Michael J. Franklin, Michael. I. Jordan, Tim Kraska
The proliferation of massive datasets combined with the development of sophisticated analytical techniques have enabled a wide variety of novel applications such as improved product recommendations, automatic image tagging, and improved speech-driven interfaces.
no code implementations • NeurIPS 2014 • Xinghao Pan, Stefanie Jegelka, Joseph E. Gonzalez, Joseph K. Bradley, Michael. I. Jordan
Many machine learning problems can be reduced to the maximization of submodular functions.
2 code implementations • 12 Sep 2014 • Daniel Crankshaw, Peter Bailis, Joseph E. Gonzalez, Haoyuan Li, Zhao Zhang, Michael J. Franklin, Ali Ghodsi, Michael. I. Jordan
In this work, we present Velox, a new component of the Berkeley Data Analytics Stack.
Databases
no code implementations • NeurIPS 2014 • Martin Jaggi, Virginia Smith, Martin Takáč, Jonathan Terhorst, Sanjay Krishnan, Thomas Hofmann, Michael. I. Jordan
Communication remains the most significant bottleneck in the performance of distributed optimization algorithms for large-scale machine learning.
no code implementations • NeurIPS 2014 • Robert Nishihara, Stefanie Jegelka, Michael. I. Jordan
Submodular functions describe a variety of discrete problems in machine learning, signal processing, and computer vision.
no code implementations • NeurIPS 2014 • Yuchen Zhang, Xi Chen, Dengyong Zhou, Michael. I. Jordan
Crowdsourcing is a popular paradigm for effectively collecting labels at low cost.
no code implementations • 5 May 2014 • John C. Duchi, Michael. I. Jordan, Martin J. Wainwright, Yuchen Zhang
Large data sets often require performing distributed statistical estimation, with a full data set split across multiple machines and limited communication between machines.
no code implementations • 3 Jan 2014 • Fredrik Lindsten, Michael. I. Jordan, Thomas B. Schön
Particle Markov chain Monte Carlo (PMCMC) is a systematic way of combining the two main tools used for Monte Carlo statistical inference: sequential Monte Carlo (SMC) and Markov chain Monte Carlo (MCMC).
no code implementations • 7 Dec 2013 • John C. Duchi, Michael. I. Jordan, Martin J. Wainwright, Andre Wibisono
We consider derivative-free algorithms for stochastic and non-stochastic convex optimization problems that use only function values rather than gradients.
no code implementations • NeurIPS 2013 • Yuchen Zhang, John Duchi, Michael. I. Jordan, Martin J. Wainwright
We establish minimax risk lower bounds for distributed statistical estimation given a budget $B$ of the total number of bits that may be communicated.
no code implementations • NeurIPS 2013 • Fabian L. Wauthier, Nebojsa Jojic, Michael. I. Jordan
The Lasso is a cornerstone of modern multivariate data analysis, yet its performance suffers in the common situation in which covariates are correlated.
no code implementations • NeurIPS 2013 • John Duchi, Martin J. Wainwright, Michael. I. Jordan
We provide a detailed study of the estimation of probability distributions---discrete and continuous---in a stringent setting in which data is kept private even from the statistician.
no code implementations • NeurIPS 2013 • John Duchi, Michael. I. Jordan, Brendan Mcmahan
We study stochastic optimization problems when the \emph{data} is sparse, which is in a sense dual to the current understanding of high-dimensional statistical learning and optimization.
no code implementations • 21 Oct 2013 • Evan R. Sparks, Ameet Talwalkar, Virginia Smith, Jey Kottalam, Xinghao Pan, Joseph Gonzalez, Michael J. Franklin, Michael. I. Jordan, Tim Kraska
MLI is an Application Programming Interface designed to address the challenges of building Machine Learn- ing algorithms in a distributed setting based on data-centric computing.
no code implementations • 30 Sep 2013 • Michael. I. Jordan
How should statistical procedures be designed so as to be scalable computationally to the massive datasets that are increasingly the norm?
no code implementations • 13 Sep 2013 • Emily B. Fox, Michael. I. Jordan
Although much of the literature on mixed membership models considers the setting in which exchangeable collections of data are associated with each member of a set of entities, it is equally natural to consider problems in which an entire time series is viewed as an entity and the goal is to characterize the time series in terms of a set of underlying dynamic attributes or "dynamic regimes".
no code implementations • 22 Aug 2013 • Emily B. Fox, Michael C. Hughes, Erik B. Sudderth, Michael. I. Jordan
We propose a Bayesian nonparametric approach to the problem of jointly modeling multiple related time series.
no code implementations • NeurIPS 2013 • Xinghao Pan, Joseph E. Gonzalez, Stefanie Jegelka, Tamara Broderick, Michael. I. Jordan
Research on distributed machine learning algorithms has focused primarily on one of two extremes - algorithms that obey strict concurrency constraints or algorithms that obey few or no such constraints.
2 code implementations • NeurIPS 2013 • Tamara Broderick, Nicholas Boyd, Andre Wibisono, Ashia C. Wilson, Michael. I. Jordan
We present SDA-Bayes, a framework for (S)treaming, (D)istributed, (A)synchronous computation of a Bayesian posterior.
no code implementations • 20 Apr 2013 • Ameet Talwalkar, Lester Mackey, Yadong Mu, Shih-Fu Chang, Michael. I. Jordan
Vision problems ranging from image clustering to motion segmentation to semi-supervised learning can naturally be framed as subspace segmentation problems, in which one aims to recover multiple low-dimensional subspaces from noisy and corrupted input data.
1 code implementation • 23 Jan 2013 • Kevin Murphy, Yair Weiss, Michael. I. Jordan
Recently, researchers have demonstrated that loopy belief propagation - the use of Pearls polytree algorithm IN a Bayesian network WITH loops OF error- correcting codes. The most dramatic instance OF this IS the near Shannon - limit performance OF Turbo Codes codes whose decoding algorithm IS equivalent TO loopy belief propagation IN a chain - structured Bayesian network.
no code implementations • 10 Jan 2013 • Nando de Freitas, Pedro Hojen-Sorensen, Michael. I. Jordan, Stuart Russell
One of these algorithms is a mixture of two MCMC kernels: a random walk Metropolis kernel and a blockMetropolis-Hastings (MH) kernel with a variational approximation as proposaldistribution.
no code implementations • NeurIPS 2012 • Ke Jiang, Brian Kulis, Michael. I. Jordan
Links between probabilistic and non-probabilistic learning algorithms can arise by performing small-variance asymptotics, i. e., letting the variance of particular distributions in a graphical model go to zero.
no code implementations • NeurIPS 2012 • Fredrik Lindsten, Thomas Schön, Michael. I. Jordan
We present a novel method in the family of particle MCMC methods that we refer to as particle Gibbs with ancestor sampling (PG-AS).
no code implementations • NeurIPS 2012 • Andre Wibisono, Martin J. Wainwright, Michael. I. Jordan, John C. Duchi
We consider derivative-free algorithms for stochastic optimization problems that use only noisy function values rather than gradients, analyzing their finite-sample convergence rates.
no code implementations • 25 Oct 2012 • John Paisley, Chong Wang, David M. Blei, Michael. I. Jordan
We develop a nested hierarchical Dirichlet process (nHDP) for hierarchical topic modeling.
no code implementations • NeurIPS 2012 • John C. Duchi, Michael. I. Jordan, Martin J. Wainwright
We study statistical risk minimization problems under a privacy model in which the data is kept confidential even from the learner.
no code implementations • 17 Sep 2012 • Barzan Mozafari, Purnamrita Sarkar, Michael J. Franklin, Michael. I. Jordan, Samuel Madden
Based on this observation, we present two new active learning algorithms to combine humans and algorithms together in a crowd-sourced database.
no code implementations • 7 Apr 2012 • John C. Duchi, Lester Mackey, Michael. I. Jordan
With these negative results as motivation, we present a new approach to supervised ranking based on aggregation of partial preferences, and we develop $U$-statistic-based empirical risk minimization procedures.
no code implementations • NeurIPS 2011 • Fabian L. Wauthier, Michael. I. Jordan
This approach can account for more complex bias patterns that arise in ambiguous or hard labeling tasks and allows us to merge data curation and learning into a single computation.
no code implementations • NeurIPS 2011 • Lester W. Mackey, Michael. I. Jordan, Ameet Talwalkar
This work introduces Divide-Factor-Combine (DFC), a parallel divide-and-conquer framework for noisy matrix factorization.
no code implementations • 8 Nov 2011 • Tamara Broderick, Lester Mackey, John Paisley, Michael. I. Jordan
We show that the NBP is conjugate to the beta process, and we characterize the posterior distribution under the beta-negative binomial process (BNBP) and hierarchical models based on the BNBP (the HBNBP).
no code implementations • 5 Jul 2011 • Lester Mackey, Ameet Talwalkar, Michael. I. Jordan
If learning methods are to scale to the massive sizes of modern datasets, it is essential for the field of machine learning to embrace parallel and distributed computing.
no code implementations • 14 Apr 2011 • Donghui Yan, Aiyou Chen, Michael. I. Jordan
The search for good local clusterings is guided by a cluster quality measure kappa.
no code implementations • NeurIPS 2010 • Ariel Kleiner, Ali Rahimi, Michael. I. Jordan
We present a novel algorithm, Random Conic Pursuit, that solves semidefinite programs (SDPs) via repeated optimization over randomly selected two-dimensional subcones of the PSD cone.
no code implementations • NeurIPS 2010 • Alexandre Bouchard-Côté, Michael. I. Jordan
Since the discovery of sophisticated fully polynomial randomized algorithms for a range of #P problems (Karzanov et al., 1991; Jerrum et al., 2001; Wilson, 2004), theoretical work on approximate inference in combinatorial spaces has focused on Markov chain Monte Carlo methods.