1 code implementation • 13 Jan 2024 • Michael Kölle, Yannick Erpelding, Fabian Ritz, Thomy Phan, Steffen Illium, Claudia Linnhoff-Popien
Recent advances in Multi-Agent Reinforcement Learning have prompted the modeling of intricate interactions between agents in simulated environments.
no code implementations • 7 Jan 2024 • Robert Müller, Hasan Turalic, Thomy Phan, Michael Kölle, Jonas Nüßlein, Claudia Linnhoff-Popien
In the realm of Multi-Agent Reinforcement Learning (MARL), prevailing approaches exhibit shortcomings in aligning with human learning, robustness, and scalability.
1 code implementation • 28 Dec 2023 • Thomy Phan, Taoan Huang, Bistra Dilkina, Sven Koenig
State-of-the-art anytime MAPF is based on Large Neighborhood Search (LNS), where a fast initial solution is iteratively optimized by destroying and repairing a fixed number of parts, i. e., the neighborhood, of the solution, using randomized destroy heuristics and prioritized planning.
1 code implementation • 18 Dec 2023 • Philipp Altmann, Jonas Stein, Michael Kölle, Adelina Bärligea, Thomas Gabor, Thomy Phan, Sebastian Feld, Claudia Linnhoff-Popien
Quantum computing (QC) in the current NISQ era is still limited in size and precision.
no code implementations • 9 Nov 2023 • Michael Kölle, Felix Topp, Thomy Phan, Philipp Altmann, Jonas Nüßlein, Claudia Linnhoff-Popien
We showed that our Variational Quantum Circuit approaches perform significantly better compared to a neural network with a similar amount of trainable parameters.
1 code implementation • 26 Apr 2023 • Philipp Altmann, Fabian Ritz, Leonard Feuchtinger, Jonas Nüßlein, Claudia Linnhoff-Popien, Thomy Phan
Current state-of-the-art approaches for generalization apply data augmentation techniques to increase the diversity of training data.
no code implementations • 18 Jan 2023 • Philipp Altmann, Thomy Phan, Fabian Ritz, Thomas Gabor, Claudia Linnhoff-Popien
We propose discriminative reward co-training (DIRECT) as an extension to deep reinforcement learning algorithms.
no code implementations • 10 Aug 2022 • Fabian Ritz, Thomy Phan, Andreas Sedlmeier, Philipp Altmann, Jan Wieghardt, Reiner Schmid, Horst Sauer, Cornel Klein, Claudia Linnhoff-Popien, Thomas Gabor
We define a comprehensive SD process model for ML that encompasses most tasks and artifacts described in the literature in a consistent way.
1 code implementation • NeurIPS 2021 • Thomy Phan, Fabian Ritz, Lenz Belzner, Philipp Altmann, Thomas Gabor, Claudia Linnhoff-Popien
We evaluate VAST in three multi-agent domains and show that VAST can significantly outperform state-of-the-art VFF, when the number of agents is sufficiently large.
1 code implementation • ALIFE 2021 • Fabian Ritz, Daniel Ratke, Thomy Phan, Lenz Belzner, Claudia Linnhoff-Popien
This paper considers sustainable and cooperative behavior in multi-agent systems.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 14 Dec 2020 • Fabian Ritz, Thomy Phan, Robert Müller, Thomas Gabor, Andreas Sedlmeier, Marc Zeller, Jan Wieghardt, Reiner Schmid, Horst Sauer, Cornel Klein, Claudia Linnhoff-Popien
A characteristic of reinforcement learning is the ability to develop unforeseen strategies when solving problems.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 9 Aug 2020 • Markus Friedrich, Sebastian Feld, Thomy Phan, Pierre-Alain Fayolle
Extracting a Construction Tree from potentially noisy point clouds is an important aspect of Reverse Engineering tasks in Computer Aided Design.
no code implementations • 29 Apr 2020 • Thomas Gabor, Leo Sünkel, Fabian Ritz, Thomy Phan, Lenz Belzner, Christoph Roch, Sebastian Feld, Claudia Linnhoff-Popien
We discuss the synergetic connection between quantum computing and artificial intelligence.
no code implementations • 29 Apr 2020 • Thomas Gabor, Sebastian Feld, Hila Safi, Thomy Phan, Claudia Linnhoff-Popien
Current hardware limitations restrict the potential when solving quadratic unconstrained binary optimization (QUBO) problems via the quantum approximate optimization algorithm (QAOA) or quantum annealing (QA).
2 code implementations • 11 Mar 2020 • Christoph Roch, Alexander Impertro, Thomy Phan, Thomas Gabor, Sebastian Feld, Claudia Linnhoff-Popien
Such algorithms are usually implemented in a variational form, combining a classical optimization method with a quantum machine to find good solutions to an optimization problem.
Quantum Physics
no code implementations • 31 Dec 2019 • Andreas Sedlmeier, Thomas Gabor, Thomy Phan, Lenz Belzner, Claudia Linnhoff-Popien
We further present a first viable solution for calculating a dynamic classification threshold, based on the uncertainty distribution of the training data.
1 code implementation • 11 Jul 2019 • Thomy Phan, Thomas Gabor, Robert Müller, Christoph Roch, Claudia Linnhoff-Popien
We propose Stable Yet Memory Bounded Open-Loop (SYMBOL) planning, a general memory bounded approach to partially observable open-loop planning.
no code implementations • 10 May 2019 • Carsten Hahn, Thomy Phan, Thomas Gabor, Lenz Belzner, Claudia Linnhoff-Popien
In nature, flocking or swarm behavior is observed in many species as it has beneficial properties like reducing the probability of being caught by a predator.
Multi-agent Reinforcement Learning reinforcement-learning +1
1 code implementation • 10 May 2019 • Thomy Phan, Lenz Belzner, Marie Kiermeier, Markus Friedrich, Kyrill Schmid, Claudia Linnhoff-Popien
State-of-the-art approaches to partially observable planning like POMCP are based on stochastic tree search.
no code implementations • 25 Jan 2019 • Thomy Phan, Kyrill Schmid, Lenz Belzner, Thomas Gabor, Sebastian Feld, Claudia Linnhoff-Popien
We experimentally evaluate STEP in two challenging and stochastic domains with large state and joint action spaces and show that STEP is able to learn stronger policies than standard multi-agent reinforcement learning algorithms, when combining multi-agent open-loop planning with centralized function approximation.
no code implementations • 8 Jan 2019 • Andreas Sedlmeier, Thomas Gabor, Thomy Phan, Lenz Belzner, Claudia Linnhoff-Popien
Although prior work has shown that dropout-based variational inference techniques and bootstrap-based approaches can be used to model epistemic uncertainty, the suitability for detecting OOD samples in deep reinforcement learning remains an open question.
no code implementations • 30 Oct 2018 • Thomas Gabor, Lenz Belzner, Thomy Phan, Kyrill Schmid
As automatic optimization techniques find their way into industrial applications, the behavior of many complex systems is determined by some form of planner picking the right actions to optimize a given objective function.