Search Results for author: Kefan Dong

Found 13 papers, 3 papers with code

Toward $L_\infty$-recovery of Nonlinear Functions: A Polynomial Sample Complexity Bound for Gaussian Random Fields

no code implementations29 Apr 2023 Kefan Dong, Tengyu Ma

Our key technical novelty is to prove that the degree-$k$ spherical harmonics components of a function from Gaussian random field cannot be spiky in that their $L_\infty$/$L_2$ ratios are upperbounded by $O(d \sqrt{\ln k})$ with high probability.

Model-based Offline Reinforcement Learning with Local Misspecification

no code implementations26 Jan 2023 Kefan Dong, Yannis Flet-Berliac, Allen Nie, Emma Brunskill

We present a model-based offline reinforcement learning policy performance lower bound that explicitly captures dynamics model misspecification and distribution mismatch and we propose an empirical algorithm for optimal offline policy selection.

D4RL reinforcement-learning +1

First Steps Toward Understanding the Extrapolation of Nonlinear Models to Unseen Domains

no code implementations21 Nov 2022 Kefan Dong, Tengyu Ma

The question is very challenging because even two-layer neural networks cannot be guaranteed to extrapolate outside the support of the training distribution without further assumptions on the domain shift.

Asymptotic Instance-Optimal Algorithms for Interactive Decision Making

no code implementations6 Jun 2022 Kefan Dong, Tengyu Ma

Past research on interactive decision making problems (bandits, reinforcement learning, etc.)

Decision Making Multi-Armed Bandits +2

Design of Experiments for Stochastic Contextual Linear Bandits

no code implementations NeurIPS 2021 Andrea Zanette, Kefan Dong, Jonathan Lee, Emma Brunskill

In the stochastic linear contextual bandit setting there exist several minimax procedures for exploration with policies that are reactive to the data being acquired.

Refined Analysis of FPL for Adversarial Markov Decision Processes

no code implementations21 Aug 2020 Yuanhao Wang, Kefan Dong

We consider the adversarial Markov Decision Process (MDP) problem, where the rewards for the MDP can be adversarially chosen, and the transition function can be either known or unknown.

Multinomial Logit Bandit with Low Switching Cost

no code implementations ICML 2020 Kefan Dong, Yingkai Li, Qin Zhang, Yuan Zhou

We also present the ESUCB algorithm with item switching cost $O(N \log^2 T)$.

On the Expressivity of Neural Networks for Deep Reinforcement Learning

1 code implementation ICML 2020 Kefan Dong, Yuping Luo, Tengyu Ma

We compare the model-free reinforcement learning with the model-based approaches through the lens of the expressive power of neural networks for policies, $Q$-functions, and dynamics.

reinforcement-learning Reinforcement Learning (RL)

Bootstrapping the Expressivity with Model-based Planning

1 code implementation25 Sep 2019 Kefan Dong, Yuping Luo, Tengyu Ma

We compare the model-free reinforcement learning with the model-based approaches through the lens of the expressive power of neural networks for policies, $Q$-functions, and dynamics.

$\sqrt{n}$-Regret for Learning in Markov Decision Processes with Function Approximation and Low Bellman Rank

no code implementations5 Sep 2019 Kefan Dong, Jian Peng, Yining Wang, Yuan Zhou

Our learning algorithm, Adaptive Value-function Elimination (AVE), is inspired by the policy elimination algorithm proposed in (Jiang et al., 2017), known as OLIVE.

Efficient Exploration

Exploration via Hindsight Goal Generation

1 code implementation NeurIPS 2019 Zhizhou Ren, Kefan Dong, Yuan Zhou, Qiang Liu, Jian Peng

Goal-oriented reinforcement learning has recently been a practical framework for robotic manipulation tasks, in which an agent is required to reach a certain goal defined by a function on the state space.

reinforcement-learning Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.