Search Results for author: Dhruv Rohatgi

Found 11 papers, 0 papers with code

Exploration is Harder than Prediction: Cryptographically Separating Reinforcement Learning from Supervised Learning

no code implementations4 Apr 2024 Noah Golowich, Ankur Moitra, Dhruv Rohatgi

We also show that there is no computationally efficient algorithm for reward-directed RL in block MDPs, even when given access to an oracle for this regression problem.

regression Reinforcement Learning (RL)

Lasso with Latents: Efficient Estimation, Covariate Rescaling, and Computational-Statistical Gaps

no code implementations23 Feb 2024 Jonathan Kelner, Frederic Koehler, Raghu Meka, Dhruv Rohatgi

It is well-known that the statistical performance of Lasso can suffer significantly when the covariates of interest have strong correlations.

regression

Exploring and Learning in Sparse Linear MDPs without Computationally Intractable Oracles

no code implementations18 Sep 2023 Noah Golowich, Ankur Moitra, Dhruv Rohatgi

The key assumption underlying linear Markov Decision Processes (MDPs) is that the learner has access to a known feature map $\phi(x, a)$ that maps state-action pairs to $d$-dimensional vectors, and that the rewards and transitions are linear functions in this representation.

feature selection Learning Theory +1

Learning in Observable POMDPs, without Computationally Intractable Oracles

no code implementations7 Jun 2022 Noah Golowich, Ankur Moitra, Dhruv Rohatgi

Much of reinforcement learning theory is built on top of oracles that are computationally hard to implement.

Learning Theory Reinforcement Learning (RL)

Provably Auditing Ordinary Least Squares in Low Dimensions

no code implementations28 May 2022 Ankur Moitra, Dhruv Rohatgi

Measuring the stability of conclusions derived from Ordinary Least Squares linear regression is critically important, but most metrics either only measure local stability (i. e. against infinitesimal changes in the data), or are only interpretable under statistical assumptions.

regression

Planning in Observable POMDPs in Quasipolynomial Time

no code implementations12 Jan 2022 Noah Golowich, Ankur Moitra, Dhruv Rohatgi

Our main result is a quasipolynomial-time algorithm for planning in (one-step) observable POMDPs.

Robust Generalized Method of Moments: A Finite Sample Viewpoint

no code implementations6 Oct 2021 Dhruv Rohatgi, Vasilis Syrgkanis

For many inference problems in statistics and econometrics, the unknown parameter is identified by a set of moment conditions.

Econometrics regression +1

On the Power of Preconditioning in Sparse Linear Regression

no code implementations17 Jun 2021 Jonathan Kelner, Frederic Koehler, Raghu Meka, Dhruv Rohatgi

First, we show that the preconditioned Lasso can solve a large class of sparse linear regression problems nearly optimally: it succeeds whenever the dependency structure of the covariates, in the sense of the Markov property, has low treewidth -- even if $\Sigma$ is highly ill-conditioned.

regression

Truncated Linear Regression in High Dimensions

no code implementations NeurIPS 2020 Constantinos Daskalakis, Dhruv Rohatgi, Manolis Zampetakis

As a corollary, our guarantees imply a computationally efficient and information-theoretically optimal algorithm for compressed sensing with truncation, which may arise from measurement saturation effects.

regression Vocal Bursts Intensity Prediction

Constant-Expansion Suffices for Compressed Sensing with Generative Priors

no code implementations NeurIPS 2020 Constantinos Daskalakis, Dhruv Rohatgi, Manolis Zampetakis

Using this theorem we can show that a matrix concentration inequality known as the Weight Distribution Condition (WDC), which was previously only known to hold for Gaussian matrices with logarithmic aspect ratio, in fact holds for constant aspect ratios too.

Retrieval

Cannot find the paper you are looking for? You can Submit a new open access paper.