Active Model Estimation in Markov Decision Processes

We study the problem of efficient exploration in order to learn an accurate model of an environment, modeled as a Markov decision process (MDP). Efficient exploration in this problem requires the agent to identify the regions in which estimating the model is more difficult and then exploit this knowledge to collect more samples there... (read more)

Results in Papers With Code
(↓ scroll down to see all results)