Search Results for author: Peter Hellinckx

Found 13 papers, 1 papers with code

Autonomous Port Navigation With Ranging Sensors Using Model-Based Reinforcement Learning

no code implementations17 Nov 2023 Siemen Herremans, Ali Anwar, Arne Troch, Ian Ravijts, Maarten Vangeneugden, Siegfried Mercelis, Peter Hellinckx

The proposed methodology is based on a machine learning approach that has recently set benchmark results in various domains: model-based reinforcement learning.

Model-based Reinforcement Learning Navigate +1

Safety Aware Autonomous Path Planning Using Model Predictive Reinforcement Learning for Inland Waterways

no code implementations16 Nov 2023 Astrid Vanneste, Simon Vanneste, Olivier Vasseur, Robin Janssens, Mattias Billast, Ali Anwar, Kevin Mets, Tom De Schepper, Siegfried Mercelis, Peter Hellinckx

We demonstrate our approach on two scenarios and compare the resulting path with path planning using a Frenet frame and path planning based on a proximal policy optimization (PPO) agent.

Navigate reinforcement-learning

Scalability of Message Encoding Techniques for Continuous Communication Learned with Multi-Agent Reinforcement Learning

no code implementations9 Aug 2023 Astrid Vanneste, Thomas Somers, Simon Vanneste, Kevin Mets, Tom De Schepper, Siegfried Mercelis, Peter Hellinckx

Therefore, we analyse the communication protocol used by the agents that use the mean message encoder and can conclude that the agents use a combination of an exponential and a logarithmic function in their communication policy to avoid the loss of important information after applying the mean message encoder.

Multi-agent Reinforcement Learning

An Analysis of Discretization Methods for Communication Learning with Multi-Agent Reinforcement Learning

no code implementations12 Apr 2022 Astrid Vanneste, Simon Vanneste, Kevin Mets, Tom De Schepper, Siegfried Mercelis, Steven Latré, Peter Hellinckx

The most common approach to allow learned communication between agents is the use of a differentiable communication channel that allows gradients to flow between agents as a form of feedback.

Multi-agent Reinforcement Learning reinforcement-learning +1

Learning to Communicate with Reinforcement Learning for an Adaptive Traffic Control System

no code implementations29 Oct 2021 Simon Vanneste, Gauthier de Borrekens, Stig Bosmans, Astrid Vanneste, Kevin Mets, Siegfried Mercelis, Steven Latré, Peter Hellinckx

In this paper, we investigate independent Q-learning (IQL) without communication and differentiable inter-agent learning (DIAL) with learned communication on an adaptive traffic control system (ATCS).

Multi-agent Reinforcement Learning Q-Learning +2

Deep Learning of Intrinsically Motivated Options in the Arcade Learning Environment

no code implementations29 Sep 2021 Louis Bagot, Kevin Mets, Tom De Schepper, Peter Hellinckx, Steven Latre

As an alternative to the widespread method of a weighted sum of rewards, Explore Options let the agent call an intrinsically motivated agent in order to observe and learn from interesting behaviors in the environment.

Atari Games Benchmarking +3

Exploiting non-i.i.d. data towards more robust machine learning algorithms

no code implementations7 Oct 2020 Wim Casteels, Peter Hellinckx

The resulting algorithm favours correlations that are universal over the subpopulations and indeed a better performance is obtained on an out-of-distribution test set with respect to a more conventional l_2-regularization.

BIG-bench Machine Learning Clustering

Learning to Communicate Using Counterfactual Reasoning

no code implementations12 Jun 2020 Simon Vanneste, Astrid Vanneste, Kevin Mets, Tom De Schepper, Ali Anwar, Siegfried Mercelis, Steven Latré, Peter Hellinckx

The credit assignment problem, the non-stationarity of the communication environment and the creation of influenceable agents are major challenges within this research field which need to be overcome in order to learn a valid communication protocol.

counterfactual Counterfactual Reasoning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.