1 code implementation • 29 Aug 2024 • Maria Laura Santoni, Elena Raponi, Aneta Neumann, Frank Neumann, Mike Preuss, Carola Doerr
We emphasize that the main goal of our work is not to present a new algorithm but to look at the problem in a more fundamental and theoretically tractable way by asking the question: What trade-off exists between the minimum distance within batches of solutions and the average quality of their fitness?
no code implementations • 19 Aug 2024 • Zhao Yang, Thomas M. Moerland, Mike Preuss, Aske Plaat, Edward S. Hu
However, the training process of RL is far from automatic, requiring extensive human effort to reset the agent and environments.
1 code implementation • 29 Sep 2023 • Marco Pleines, Matthias Pallasch, Frank Zimmer, Mike Preuss
Memory Gym presents a suite of 2D partially observable environments, namely Mortar Mayhem, Mystery Path, and Searing Spotlights, designed to benchmark memory capabilities in decision-making agents.
no code implementations • 19 Sep 2023 • Arthur van der Staaij, Jelmer Prins, Vincent L. Prins, Julian Poelsma, Thera Smit, Matthias Müller-Brockhausen, Mike Preuss
Procedural city generation that focuses on believability and adaptability to random terrain is a difficult challenge in the field of Procedural Content Generation (PCG).
no code implementations • 10 Aug 2023 • Paula Torren-Peraire, Alan Kai Hassen, Samuel Genheden, Jonas Verhoeven, Djork-Arne Clevert, Mike Preuss, Igor Tetko
Furthermore, we show that the commonly used single-step retrosynthesis benchmark dataset USPTO-50k is insufficient as this evaluation task does not represent model performance and scalability on larger and more diverse datasets.
no code implementations • 20 Apr 2023 • Zhao Yang, Thomas. M. Moerland, Mike Preuss, Aske Plaat
While deep reinforcement learning has shown important empirical success, it tends to learn relatively slow due to slow propagation of rewards information and slow update of parametric neural networks.
no code implementations • 12 Dec 2022 • Alan Kai Hassen, Paula Torren-Peraire, Samuel Genheden, Jonas Verhoeven, Mike Preuss, Igor Tetko
Retrosynthesis is the task of breaking down a chemical compound recursively step-by-step into molecular precursors until a set of commercially available molecules is found.
no code implementations • 6 Dec 2022 • Zhao Yang, Thomas M. Moerland, Mike Preuss, Aske Plaat
In this paper, we present a clear ablation study of post-exploration in a general intrinsically motivated goal exploration process (IMGEP) framework, that the Go-Explore paper did not show.
no code implementations • 28 Nov 2022 • Zhao Yang, Thomas M. Moerland, Mike Preuss, Aske Plaat
Therefore, this paper introduces Continuous Episodic Control (CEC), a novel non-parametric episodic memory algorithm for sequential decision making in problems with a continuous action space.
no code implementations • 23 May 2022 • Marco Pleines, Matthias Pallasch, Frank Zimmer, Mike Preuss
At first sight it may seem straightforward to use recurrent layers in Deep Reinforcement Learning algorithms to enable agents to make use of memory in the setting of partially observable environments.
no code implementations • 10 May 2022 • Marco Pleines, Konstantin Ramthun, Yannik Wegener, Hendrik Meyer, Matthias Pallasch, Sebastian Prior, Jannik Drögemüller, Leon Büttinghaus, Thilo Röthemeyer, Alexander Kaschwig, Oliver Chmurzynski, Frederik Rohkrähmer, Roman Kalkreuth, Frank Zimmer, Mike Preuss
Autonomously trained agents that are supposed to play video games reasonably well rely either on fast simulation speeds or heavy parallelization across thousands of machines running concurrently.
no code implementations • 29 Mar 2022 • Zhao Yang, Thomas M. Moerland, Mike Preuss, Aske Plaat
Go-Explore achieved breakthrough performance on challenging reinforcement learning (RL) tasks with sparse rewards.
no code implementations • 2 Mar 2022 • Matthias Müller-Brockhausen, Aske Plaat, Mike Preuss
Reinforcement Learning (RL) is one of the most dynamic research areas in Game AI and AI as a whole, and a wide variety of games are used as its prominent test problems.
no code implementations • 10 Sep 2021 • Zhao Yang, Mike Preuss, Aske Plaat
While previous work has investigated the use of expert knowledge to generate potential functions, in this work, we study whether we can use a search algorithm(A*) to automatically generate a potential function for reward shaping in Sokoban, a well-known planning task.
no code implementations • 17 Jul 2021 • Aske Plaat, Walter Kosters, Mike Preuss
Deep reinforcement learning has shown remarkable success in the past few years.
no code implementations • 31 May 2021 • Matthias Müller-Brockhausen, Mike Preuss, Aske Plaat
We note a surprisingly late adoption of deep learning that starts in 2018.
no code implementations • 25 May 2021 • Zhao Yang, Mike Preuss, Aske Plaat
In reinforcement learning, learning actions for a behavior policy that can be applied to new environments is still a challenge, especially for tasks that involve much planning.
no code implementations • 13 May 2021 • Hui Wang, Mike Preuss, Aske Plaat
AlphaZero has achieved impressive performance in deep reinforcement learning by utilizing an architecture that combines search and training of a neural network in self-play.
no code implementations • 10 May 2021 • Alexander Hagg, Mike Preuss, Alexander Asteroth, Thomas Bäck
More and more, optimization methods are used to find diverse solution sets.
no code implementations • 25 Aug 2020 • Christoph Salge, Emily Short, Mike Preuss, Spyridion Samothrakis, Pieter Spronck
Live Action Role-Playing (LARP) games and similar experiences are becoming a popular game genre.
no code implementations • 11 Aug 2020 • Aske Plaat, Walter Kosters, Mike Preuss
In recent years, many model-based methods have been introduced to address this challenge.
no code implementations • 14 Jun 2020 • Hui Wang, Mike Preuss, Michael Emmerich, Aske Plaat
A later algorithm, Nested Rollout Policy Adaptation, was able to find a new record of 82 steps, albeit with large computational resources.
no code implementations • 29 Apr 2020 • Jialin Liu, Antoine Moreau, Mike Preuss, Baptiste Roziere, Jeremy Rapin, Fabien Teytaud, Olivier Teytaud
Choosing automatically the right algorithm using problem descriptors is a classical component of combinatorial optimization.
no code implementations • 26 Apr 2020 • Hui Wang, Mike Preuss, Aske Plaat
Recently, AlphaZero has achieved landmark results in deep reinforcement learning, by providing a single self-play architecture that learned three different games at super human level.
no code implementations • 1 Apr 2020 • Matthias Muller-Brockhausen, Mike Preuss, Aske Plaat
This paper focuses on a new game, Tetris Link, a board game that is still lacking any scientific analysis.
1 code implementation • 1 Apr 2020 • Marco Pleines, Jenia Jitsev, Mike Preuss, Frank Zimmer
The Obstacle Tower Challenge is the task to master a procedurally generated chain of levels that subsequently get harder to complete.
no code implementations • 12 Mar 2020 • Hui Wang, Michael Emmerich, Mike Preuss, Aske Plaat
A secondary result of our experiments concerns the choice of optimization goals, for which we also provide recommendations.
no code implementations • 24 Feb 2020 • Sebastian Risi, Mike Preuss
This paper reviews the field of Game AI, which not only deals with creating agents that can play a certain game, but also with areas as diverse as creating game content automatically, game analytics, or player modelling.
1 code implementation • 19 Mar 2019 • Hui Wang, Michael Emmerich, Mike Preuss, Aske Plaat
Therefore, in this paper, we choose 12 parameters in AlphaZero and evaluate how these parameters contribute to training.
no code implementations • 14 Aug 2017 • Marwin H. S. Segler, Mike Preuss, Mark P. Waller
We anticipate that our method will accelerate drug and materials discovery by assisting chemists to plan better syntheses faster, and by enabling fully automated robot synthesis.
no code implementations • 19 Apr 2017 • Simon Wessing, Mike Preuss
Efficient global optimization is a popular algorithm for the optimization of expensive multimodal black-box functions.