no code implementations • 17 Jun 2024 • Noah Golowich, Ankur Moitra
Our main structural assumption is that the MDP has low inherent Bellman error, which stipulates that linear value functions have linear Bellman backups with respect to the greedy policy.
no code implementations • 17 Jun 2024 • Noah Golowich, Ankur Moitra
One of the most natural approaches to reinforcement learning (RL) with function approximation is value iteration, which inductively generates approximations to the optimal value function by solving a sequence of regression problems.
no code implementations • 4 Jun 2024 • Noah Golowich, Ankur Moitra
We aim for watermarks which satisfy: (a) undetectability, a cryptographic notion introduced by Christ, Gunn & Zamir (2024) which stipulates that it is computationally hard to distinguish watermarked language model outputs from the model's actual output distribution; and (b) robustness to channels which introduce a constant fraction of adversarial insertions, substitutions, and deletions to the watermarked text.
no code implementations • 30 Apr 2024 • Ainesh Bakshi, Allen Liu, Ankur Moitra, Ewin Tang
We study the problem of Hamiltonian structure learning from real-time evolution: given the ability to apply $e^{-\mathrm{i} Ht}$ for an unknown local Hamiltonian $H = \sum_{a = 1}^m \lambda_a E_a$ on $n$ qubits, the goal is to recover $H$.
no code implementations • 4 Apr 2024 • Noah Golowich, Ankur Moitra, Dhruv Rohatgi
We also show that there is no computationally efficient algorithm for reward-directed RL in block MDPs, even when given access to an oracle for this regression problem.
no code implementations • 1 Nov 2023 • Ankur Moitra, Alexander S. Wein
Our result shows that the spectrum is a sufficient statistic for computationally bounded tests (but not for all tests).
no code implementations • 3 Oct 2023 • Ainesh Bakshi, Allen Liu, Ankur Moitra, Ewin Tang
Anshu, Arunachalam, Kuwahara, and Soleimanifar (arXiv:2004. 07266) gave an algorithm to learn a Hamiltonian on $n$ qubits to precision $\epsilon$ with only polynomially many copies of the Gibbs state, but which takes exponential time.
no code implementations • 18 Sep 2023 • Noah Golowich, Ankur Moitra, Dhruv Rohatgi
The key assumption underlying linear Markov Decision Processes (MDPs) is that the learner has access to a known feature map $\phi(x, a)$ that maps state-action pairs to $d$-dimensional vectors, and that the rewards and transitions are linear functions in this representation.
no code implementations • 13 Jul 2023 • Ainesh Bakshi, Allen Liu, Ankur Moitra, Morris Yau
In this work we give a new approach to learning mixtures of linear dynamical systems that is based on tensor decompositions.
no code implementations • 23 Jan 2023 • Ainesh Bakshi, Allen Liu, Ankur Moitra, Morris Yau
Linear dynamical systems are the foundational statistical model upon which control theory is built.
no code implementations • 25 Jul 2022 • Allen Liu, Ankur Moitra
In this work, we study the problem of community detection in the stochastic block model with adversarial node corruptions.
1 code implementation • 29 Jun 2022 • Saachi Jain, Hannah Lawrence, Ankur Moitra, Aleksander Madry
Moreover, by combining our framework with off-the-shelf diffusion models, we can generate images that are especially challenging for the analyzed model, and thus can be used to perform synthetic data augmentation that helps remedy the model's failure modes.
no code implementations • 7 Jun 2022 • Noah Golowich, Ankur Moitra, Dhruv Rohatgi
Much of reinforcement learning theory is built on top of oracles that are computationally hard to implement.
no code implementations • 28 May 2022 • Ankur Moitra, Dhruv Rohatgi
Measuring the stability of conclusions derived from Ordinary Least Squares linear regression is critically important, but most metrics either only measure local stability (i. e. against infinitesimal changes in the data), or are only interpretable under statistical assumptions.
no code implementations • 12 Jan 2022 • Noah Golowich, Ankur Moitra, Dhruv Rohatgi
Our main result is a quasipolynomial-time algorithm for planning in (one-step) observable POMDPs.
no code implementations • 13 Dec 2021 • Allen Liu, Ankur Moitra
Maximum likelihood estimation furnishes powerful insights into voting theory, and the design of voting rules.
no code implementations • NeurIPS 2021 • Linus Hamilton, Ankur Moitra
In recent years there has been significant effort to adapt the key tools and ideas in convex optimization to the Riemannian setting.
no code implementations • 11 Nov 2021 • Sitan Chen, Frederic Koehler, Ankur Moitra, Morris Yau
In a pioneering work, Schick and Mitter gave provable guarantees when the measurement noise is a known infinitesimal perturbation of a Gaussian and raised the important question of whether one can get similar guarantees for large and unknown perturbations.
no code implementations • 25 Oct 2021 • Noah Golowich, Ankur Moitra
In this paper we address the question of whether worst-case lower bounds for regret in online learning of Markov decision processes (MDPs) can be circumvented when information about the MDP, in the form of predictions about its optimal $Q$-value function, is given to the algorithm.
no code implementations • 29 Sep 2021 • Hannah Lawrence, Ankur Moitra
There is a rich literature on recovering data from limited measurements under the assumption of sparsity in some basis, whether known (compressed sensing) or unknown (dictionary learning).
no code implementations • 15 Jun 2021 • Ankur Moitra, Elchanan Mossel, Colin Sandon
In this work, we study the computational complexity of determining whether a machine learning model that perfectly fits the training data will generalizes to unseen data.
no code implementations • 5 Jun 2021 • Jerry Li, Allen Liu, Ankur Moitra
Given $\textsf{poly}(k/\epsilon)$ samples from a distribution that is $\epsilon$-close in TV distance to a GMM with $k$ components, we can construct a GMM with $\widetilde{O}(k)$ components that approximates the distribution to within $\widetilde{O}(\epsilon)$ in $\textsf{poly}(k/\epsilon)$ time.
no code implementations • 4 Jun 2021 • Allen Liu, Ankur Moitra
Our main result is a quasi-polynomial time algorithm for orbit recovery over $SO(3)$ in this model.
no code implementations • 19 Apr 2021 • Allen Liu, Ankur Moitra
In this work we solve the problem of robustly learning a high-dimensional Gaussian mixture model with $k$ components from $\epsilon$-corrupted samples up to accuracy $\widetilde{O}(\epsilon)$ in total variation distance for any constant $k$ and with mild assumptions on the mixture.
no code implementations • 15 Jan 2021 • Ankur Moitra, Elchanan Mossel, Colin Sandon
These are Markov Random Fields where some of the nodes are censored (not observed).
no code implementations • 14 Jan 2021 • Linus Hamilton, Ankur Moitra
In recent years there has been significant effort to adapt the key tools and ideas in convex optimization to the Riemannian setting.
no code implementations • NeurIPS 2020 • Sitan Chen, Frederic Koehler, Ankur Moitra, Morris Yau
In this paper, we revisit the problem of distribution-independently learning halfspaces under Massart noise with rate $\eta$.
no code implementations • 6 Nov 2020 • Allen Liu, Ankur Moitra
This work represents a natural coalescence of two important lines of work: learning mixtures of Gaussians and algorithmic robust statistics.
no code implementations • 8 Oct 2020 • Sitan Chen, Frederic Koehler, Ankur Moitra, Morris Yau
Our approach is based on a novel alternating minimization scheme that interleaves ordinary least-squares with a simple convex program that finds the optimal reweighting of the distribution under a spectral constraint.
1 code implementation • 8 Jun 2020 • Sitan Chen, Frederic Koehler, Ankur Moitra, Morris Yau
In particular, we study the problem of learning halfspaces under Massart noise with rate $\eta$.
no code implementations • NeurIPS 2020 • Allen Liu, Ankur Moitra
We show strong provable guarantees, including showing that our algorithm converges linearly to the true tensors even when the factors are highly correlated and can be implemented in nearly linear time.
no code implementations • ICLR Workshop DeepDiffEq 2019 • Ankur Moitra, Andrej Risteski
In this paper, we study one aspect of nonconvexity relevant for modern machine learning applications: existence of invariances (symmetries) in the function f, as a result of which the distribution p will have manifolds of points with equal probability.
1 code implementation • NeurIPS 2020 • Sitan Chen, Jerry Li, Ankur Moitra
We revisit the problem of learning from untrusted batches introduced by Qiao and Valiant [QV17].
no code implementations • 13 Feb 2020 • Ankur Moitra, Andrej Risteski
In this paper, we focus on an aspect of nonconvexity relevant for modern machine learning applications: existence of invariances (symmetries) in the function f, as a result of which the distribution p will have manifolds of points with equal probability.
no code implementations • 3 Dec 2019 • Diego Cifuentes, Ankur Moitra
The basic idea is to solve a nonconvex program in $Y$, where $Y$ is an $n \times p$ matrix such that $X = Y Y^T$.
no code implementations • 5 Nov 2019 • Sitan Chen, Jerry Li, Ankur Moitra
When $k = 1$ this is the standard robust univariate density estimation setting and it is well-understood that $\Omega (\epsilon)$ error is unavoidable.
no code implementations • NeurIPS 2020 • Jonathan Kelner, Frederic Koehler, Raghu Meka, Ankur Moitra
While there are a variety of algorithms (e. g. Graphical Lasso, CLIME) that provably recover the graph structure with a logarithmic number of samples, they assume various conditions that require the precision matrix to be in some sense well-conditioned.
no code implementations • 2 Nov 2018 • Ankur Moitra, Alexander S. Wein
Many existing algorithms for tensor problems (such as tensor decomposition and tensor PCA), although they are not presented this way, can be viewed as spectral methods on matrices built from simple tensor networks.
no code implementations • 17 Aug 2018 • Allen Liu, Ankur Moitra
Mixtures of Mallows models are a popular generative model for ranking data coming from a heterogeneous population.
no code implementations • 2 Jul 2018 • Amelia Perry, Alexander S. Wein, Afonso S. Bandeira, Ankur Moitra
Our results leverage Le Cam's notion of contiguity, and include: i) For the Gaussian Wigner ensemble, we show that PCA achieves the optimal detection threshold for certain natural priors for the spike.
no code implementations • 25 May 2018 • Guy Bresler, Frederic Koehler, Ankur Moitra, Elchanan Mossel
This hardness result is based on a sharp and surprising characterization of the representational power of bounded degree RBMs: the distribution on their observed variables can simulate any bounded order MRF.
no code implementations • 17 Mar 2018 • Sitan Chen, Ankur Moitra
In contrast, as we will show, mixtures of $k$ subcubes are uniquely determined by their degree $2 \log k$ moments and hence provide a useful abstraction for simultaneously achieving the polynomial dependence on $1/\epsilon$ of the classic Occam algorithms for decision trees and the flexibility of the low-degree algorithm in being able to accommodate stochastic transitions.
no code implementations • ICML 2017 • John Urschel, Victor-Emmanuel Brunel, Ankur Moitra, Philippe Rigollet
Determinantal Point Processes (DPPs) are a family of probabilistic models that have a repulsive behavior, and lend themselves naturally to many tasks in machine learning where returning a diverse set of objects is important.
no code implementations • NeurIPS 2017 • Linus Hamilton, Frederic Koehler, Ankur Moitra
As an application, we obtain algorithms for learning Markov random fields on bounded degree graphs on $n$ nodes with $r$-order interactions in $n^r$ time and $\log n$ sample complexity.
no code implementations • 12 Apr 2017 • Ilias Diakonikolas, Gautam Kamath, Daniel M. Kane, Jerry Li, Ankur Moitra, Alistair Stewart
We give robust estimators that achieve estimation error $O(\varepsilon)$ in the total variation distance, which is optimal up to a universal constant that is independent of the dimension.
2 code implementations • ICML 2017 • Ilias Diakonikolas, Gautam Kamath, Daniel M. Kane, Jerry Li, Ankur Moitra, Alistair Stewart
Robust estimation is much more challenging in high dimensions than it is in one dimension: Most techniques either lead to intractable optimization problems or estimators that can tolerate only a tiny fraction of errors.
no code implementations • 14 Oct 2016 • Ankur Moitra
In this paper we introduce a new approach for approximately counting in bounded degree systems with higher-order constraints.
no code implementations • 14 Oct 2016 • Amelia Perry, Alexander S. Wein, Afonso S. Bandeira, Ankur Moitra
Various alignment problems arising in cryo-electron microscopy, community detection, time synchronization, computer vision, and other fields fall into a common framework of synchronization problems over compact groups such as Z/L, U(1), or SO(3).
no code implementations • 19 Sep 2016 • Amelia Perry, Alexander S. Wein, Afonso S. Bandeira, Ankur Moitra
Our results include: I) For the Gaussian Wigner ensemble, we show that PCA achieves the optimal detection threshold for a variety of benign priors for the spike.
no code implementations • 27 May 2016 • Sanjeev Arora, Rong Ge, Frederic Koehler, Tengyu Ma, Ankur Moitra
But designing provable algorithms for inference has proven to be more challenging.
2 code implementations • 21 Apr 2016 • Ilias Diakonikolas, Gautam Kamath, Daniel Kane, Jerry Li, Ankur Moitra, Alistair Stewart
We study high-dimensional distribution learning in an agnostic setting where an adversary is allowed to arbitrarily corrupt an $\varepsilon$-fraction of the samples.
no code implementations • 4 Nov 2015 • Ankur Moitra, William Perry, Alexander S. Wein
The stochastic block model is one of the oldest and most ubiquitous models for studying clustering and community detection.
no code implementations • 2 Mar 2015 • Sanjeev Arora, Rong Ge, Tengyu Ma, Ankur Moitra
Its standard formulation is as a non-convex optimization problem which is solved in practice by heuristics based on alternating minimization.
no code implementations • 26 Jan 2015 • Boaz Barak, Ankur Moitra
This is also the first algorithm for tensor completion that works in the overcomplete case when $r > n$, and in fact it works all the way up to $r = n^{3/2-\epsilon}$.
no code implementations • 14 Nov 2013 • Aditya Bhaskara, Moses Charikar, Ankur Moitra, Aravindan Vijayaraghavan
We introduce a smoothed analysis model for studying these questions and develop an efficient algorithm for tensor decomposition in the highly overcomplete case (rank polynomial in the dimension).
no code implementations • 28 Aug 2013 • Sanjeev Arora, Rong Ge, Ankur Moitra
In sparse recovery we are given a matrix $A$ (the dictionary) and a vector of the form $A X$ where $X$ is sparse, and the goal is to recover $X$.
no code implementations • 6 Feb 2013 • Ankur Moitra, Michael Saks
This improves on algorithm of Wigderson and Yehudayoff that runs in quasi-polynomial time for any $\mu > 0$ and the polynomial time algorithm of Dvir et al which was shown to work for $\mu \gtrapprox 0. 30$ by Batman et al.
2 code implementations • 19 Dec 2012 • Sanjeev Arora, Rong Ge, Yoni Halpern, David Mimno, Ankur Moitra, David Sontag, Yichen Wu, Michael Zhu
Topic models provide a useful method for dimensionality reduction and exploratory data analysis in large text corpora.
no code implementations • NeurIPS 2012 • Sanjeev Arora, Rong Ge, Ankur Moitra, Sushant Sachdeva
We present a new algorithm for Independent Component Analysis (ICA) which has provable performance guarantees.
no code implementations • 5 Nov 2012 • Moritz Hardt, Ankur Moitra
We give an algorithm that finds $T$ when it contains more than a $\frac{d}{n}$ fraction of the points.
2 code implementations • 9 Apr 2012 • Sanjeev Arora, Rong Ge, Ankur Moitra
Topic Modeling is an approach used for automatic comprehension and classification of data in a variety of settings, and perhaps the canonical application is in uncovering thematic structure in a corpus of documents.