no code implementations • 4 Apr 2024 • Noah Golowich, Ankur Moitra, Dhruv Rohatgi
We also show that there is no computationally efficient algorithm for reward-directed RL in block MDPs, even when given access to an oracle for this regression problem.
no code implementations • 23 Feb 2024 • Jonathan Kelner, Frederic Koehler, Raghu Meka, Dhruv Rohatgi
It is well-known that the statistical performance of Lasso can suffer significantly when the covariates of interest have strong correlations.
no code implementations • 18 Sep 2023 • Noah Golowich, Ankur Moitra, Dhruv Rohatgi
The key assumption underlying linear Markov Decision Processes (MDPs) is that the learner has access to a known feature map $\phi(x, a)$ that maps state-action pairs to $d$-dimensional vectors, and that the rewards and transitions are linear functions in this representation.
no code implementations • 7 Jun 2022 • Noah Golowich, Ankur Moitra, Dhruv Rohatgi
Much of reinforcement learning theory is built on top of oracles that are computationally hard to implement.
no code implementations • 28 May 2022 • Ankur Moitra, Dhruv Rohatgi
Measuring the stability of conclusions derived from Ordinary Least Squares linear regression is critically important, but most metrics either only measure local stability (i. e. against infinitesimal changes in the data), or are only interpretable under statistical assumptions.
no code implementations • 5 Mar 2022 • Jonathan A. Kelner, Frederic Koehler, Raghu Meka, Dhruv Rohatgi
Surprisingly, at the heart of our lower bound is a new positive result in compressed sensing.
no code implementations • 12 Jan 2022 • Noah Golowich, Ankur Moitra, Dhruv Rohatgi
Our main result is a quasipolynomial-time algorithm for planning in (one-step) observable POMDPs.
no code implementations • 6 Oct 2021 • Dhruv Rohatgi, Vasilis Syrgkanis
For many inference problems in statistics and econometrics, the unknown parameter is identified by a set of moment conditions.
no code implementations • 17 Jun 2021 • Jonathan Kelner, Frederic Koehler, Raghu Meka, Dhruv Rohatgi
First, we show that the preconditioned Lasso can solve a large class of sparse linear regression problems nearly optimally: it succeeds whenever the dependency structure of the covariates, in the sense of the Markov property, has low treewidth -- even if $\Sigma$ is highly ill-conditioned.
no code implementations • NeurIPS 2020 • Constantinos Daskalakis, Dhruv Rohatgi, Manolis Zampetakis
As a corollary, our guarantees imply a computationally efficient and information-theoretically optimal algorithm for compressed sensing with truncation, which may arise from measurement saturation effects.
no code implementations • NeurIPS 2020 • Constantinos Daskalakis, Dhruv Rohatgi, Manolis Zampetakis
Using this theorem we can show that a matrix concentration inequality known as the Weight Distribution Condition (WDC), which was previously only known to hold for Gaussian matrices with logarithmic aspect ratio, in fact holds for constant aspect ratios too.