Atari Games
277 papers with code • 64 benchmarks • 6 datasets
The Atari 2600 Games task (and dataset) involves training an agent to achieve high game scores.
( Image credit: Playing Atari with Deep Reinforcement Learning )
Libraries
Use these libraries to find Atari Games models and implementationsDatasets
Latest papers
Think Before You Act: Decision Transformers with Internal Working Memory
We argue that this inefficiency stems from the forgetting phenomenon, in which a model memorizes its behaviors in parameters throughout training.
Train a Real-world Local Path Planner in One Hour via Partially Decoupled Reinforcement Learning and Vectorized Diversity
Meanwhile, the Sparrow simulator utilizes a 2D grid-based world, simplified kinematics, and conversion-free data flow to achieve a lightweight design.
Proto-Value Networks: Scaling Representation Learning with Auxiliary Tasks
Combined with a suitable off-policy learning rule, the result is a representation learning algorithm that can be understood as extending Mahadevan & Maggioni (2007)'s proto-value functions to deep reinforcement learning -- accordingly, we call the resulting object proto-value networks.
Unsupervised Representation Learning in Partially Observable Atari Games
Contrastive methods have performed better than generative models in previous state representation learning research.
How To Guide Your Learner: Imitation Learning with Active Adaptive Expert Involvement
In this paper, we propose a novel active imitation learning framework based on a teacher-student interaction model, in which the teacher's goal is to identify the best teaching behavior and actively affect the student's learning process.
Self-supervised network distillation: an effective approach to exploration in sparse reward environments
The solution to such a problem may be to equip the agent with an intrinsic motivation that will provide informed exploration during which the agent is likely to also encounter external reward.
Revisiting Bellman Errors for Offline Model Selection
Offline model selection (OMS), that is, choosing the best policy from a set of many policies given only logged data, is crucial for applying offline RL in real-world settings.
Scalable Real-Time Recurrent Learning Using Columnar-Constructive Networks
We show that by either decomposing the network into independent modules or learning the network in stages, we can make RTRL scale linearly with the number of parameters.
Learning to Perceive in Deep Model-Free Reinforcement Learning
We investigate whether a model with these characteristics is capable of achieving similar performance to state-of-the-art model-free RL agents that access the full input observation.
Boosting Object Representation Learning via Motion and Object Continuity
Recent unsupervised multi-object detection models have shown impressive performance improvements, largely attributed to novel architectural inductive biases.