Unsupervised Reinforcement Learning
26 papers with code • 8 benchmarks • 2 datasets
Libraries
Use these libraries to find Unsupervised Reinforcement Learning models and implementationsMost implemented papers
Exploration by Random Network Distillation
In particular we establish state of the art performance on Montezuma's Revenge, a game famously difficult for deep reinforcement learning methods.
Curiosity-driven Exploration by Self-supervised Prediction
In many real-world scenarios, rewards extrinsic to the agent are extremely sparse, or absent altogether.
Mastering Visual Continuous Control: Improved Data-Augmented Reinforcement Learning
We present DrQ-v2, a model-free reinforcement learning (RL) algorithm for visual continuous control.
Diversity is All You Need: Learning Skills without a Reward Function
On a variety of simulated robotic tasks, we show that this simple objective results in the unsupervised emergence of diverse skills, such as walking and jumping.
Self-Supervised Exploration via Disagreement
In this paper, we propose a formulation for exploration inspired by the work in active learning literature.
Emergent Real-World Robotic Skills via Unsupervised Off-Policy Reinforcement Learning
Can we instead develop efficient reinforcement learning methods that acquire diverse skills without any reward function, and then repurpose these skills for downstream tasks?
Unsupervised Reinforcement Learning in Multiple Environments
Along this line, we address the problem of unsupervised reinforcement learning in a class of multiple environments, in which the policy is pre-trained with interactions from the whole class, and then fine-tuned for several tasks in any environment of the class.
Variational Intrinsic Control
In this paper we introduce a new unsupervised reinforcement learning method for discovering the set of intrinsic options available to an agent.
Efficient Exploration via State Marginal Matching
The SMM objective can be viewed as a two-player, zero-sum game between a state density model and a parametric policy, an idea that we use to build an algorithm for optimizing the SMM objective.
SMiRL: Surprise Minimizing Reinforcement Learning in Unstable Environments
Every living organism struggles against disruptive environmental forces to carve out and maintain an orderly niche.