Search Results for author: Marco Pleines

Found 5 papers, 2 papers with code

Memory Gym: Towards Endless Tasks to Benchmark Memory Capabilities of Agents

1 code implementation29 Sep 2023 Marco Pleines, Matthias Pallasch, Frank Zimmer, Mike Preuss

Memory Gym presents a suite of 2D partially observable environments, namely Mortar Mayhem, Mystery Path, and Searing Spotlights, designed to benchmark memory capabilities in decision-making agents.

Decision Making

Improving Bidding and Playing Strategies in the Trick-Taking game Wizard using Deep Q-Networks

no code implementations27 May 2022 Jonas Schumacher, Marco Pleines

In this work, the trick-taking game Wizard with a separate bidding and playing phase is modeled by two interleaved partially observable Markov decision processes (POMDP).

Decision Making

Generalization, Mayhems and Limits in Recurrent Proximal Policy Optimization

no code implementations23 May 2022 Marco Pleines, Matthias Pallasch, Frank Zimmer, Mike Preuss

At first sight it may seem straightforward to use recurrent layers in Deep Reinforcement Learning algorithms to enable agents to make use of memory in the setting of partially observable environments.

Benchmarking

Obstacle Tower Without Human Demonstrations: How Far a Deep Feed-Forward Network Goes with Reinforcement Learning

1 code implementation1 Apr 2020 Marco Pleines, Jenia Jitsev, Mike Preuss, Frank Zimmer

The Obstacle Tower Challenge is the task to master a procedurally generated chain of levels that subsequently get harder to complete.

Cannot find the paper you are looking for? You can Submit a new open access paper.