TASK | DATASET | MODEL | METRIC NAME | METRIC VALUE | GLOBAL RANK | REMOVE |
---|---|---|---|---|---|---|
Atari Games | Atari 2600 Beam Rider | DQN Best | Score | 5184 | # 34 |
|
Atari Games | Atari 2600 Breakout | DQN Best | Score | 225 | # 36 |
|
Atari Games | Atari 2600 Enduro | DQN Best | Score | 661 | # 25 |
|
Atari Games | Atari 2600 Pong | DQN Best | Score | 21 | # 1 |
|
Atari Games | Atari 2600 Q*Bert | DQN Best | Score | 4500 | # 35 |
|
Atari Games | Atari 2600 Seaquest | DQN Best | Score | 1740 | # 33 |
|
Atari Games | Atari 2600 Space Invaders | DQN Best | Score | 1075 | # 36 |
|
Include the markdown at the top of your
GitHub README.md
file to
showcase the performance of the model.
Badges are live and will be dynamically updated with the latest ranking of this paper.
|
Markdown |
---|---|
|
[](https://paperswithcode.com/sota/atari-games-on-atari-2600-pong?p=playing-atari-with-deep-reinforcement)
|
|
[](https://paperswithcode.com/sota/atari-games-on-atari-2600-enduro?p=playing-atari-with-deep-reinforcement)
|
|
[](https://paperswithcode.com/sota/atari-games-on-atari-2600-seaquest?p=playing-atari-with-deep-reinforcement)
|
|
[](https://paperswithcode.com/sota/atari-games-on-atari-2600-beam-rider?p=playing-atari-with-deep-reinforcement)
|
|
[](https://paperswithcode.com/sota/atari-games-on-atari-2600-qbert?p=playing-atari-with-deep-reinforcement)
|
|
[](https://paperswithcode.com/sota/atari-games-on-atari-2600-breakout?p=playing-atari-with-deep-reinforcement)
|
|
[](https://paperswithcode.com/sota/atari-games-on-atari-2600-space-invaders?p=playing-atari-with-deep-reinforcement)
|
19 Dec 2013 • Volodymyr Mnih • Koray Kavukcuoglu • David Silver • Alex Graves • Ioannis Antonoglou • Daan Wierstra • Martin Riedmiller
We present the first deep learning model to successfully learn control policies directly from high-dimensional sensory input using reinforcement learning. The model is a convolutional neural network, trained with a variant of Q-learning, whose input is raw pixels and whose output is a value function estimating future rewards... We apply our method to seven Atari 2600 games from the Arcade Learning Environment, with no adjustment of the architecture or learning algorithm. We find that it outperforms all previous approaches on six of the games and surpasses a human expert on three of them. (read more)
PDF