Search Results for author: Nicolas Drougard

Found 7 papers, 3 papers with code

Offline Risk-sensitive RL with Partial Observability to Enhance Performance in Human-Robot Teaming

no code implementations8 Feb 2024 Giorgio Angelotti, Caroline P. C. Chanel, Adam H. M. Pinto, Christophe Lounis, Corentin Chauffaut, Nicolas Drougard

The integration of physiological computing into mixed-initiative human-robot interaction systems offers valuable advantages in autonomous task allocation by incorporating real-time features as human state observations into the decision-making system.

Decision Making Physiological Computing

Data Augmentation through Expert-guided Symmetry Detection to Improve Performance in Offline Reinforcement Learning

1 code implementation18 Dec 2021 Giorgio Angelotti, Nicolas Drougard, Caroline P. C. Chanel

Offline estimation of the dynamical model of a Markov Decision Process (MDP) is a non-trivial task that greatly depends on the data available in the learning phase.

Data Augmentation Density Estimation +3

Expert-Guided Symmetry Detection in Markov Decision Processes

no code implementations19 Nov 2021 Giorgio Angelotti, Nicolas Drougard, Caroline P. C. Chanel

Learning a Markov Decision Process (MDP) from a fixed batch of trajectories is a non-trivial task whose outcome's quality depends on both the amount and the diversity of the sampled regions of the state-action space.

Density Estimation Symmetry Detection

An Offline Risk-aware Policy Selection Method for Bayesian Markov Decision Processes

1 code implementation27 May 2021 Giorgio Angelotti, Nicolas Drougard, Caroline Ponzoni Carvalho Chanel

In Offline Model Learning for Planning and in Offline Reinforcement Learning, the limited data set hinders the estimate of the Value function of the relative Markov Decision Process (MDP).

reinforcement-learning

Offline Learning for Planning: A Summary

no code implementations5 Oct 2020 Giorgio Angelotti, Nicolas Drougard, Caroline Ponzoni Carvalho Chanel

Offline learning is the area of machine learning concerned with efficiently obtaining an optimal policy with a batch of previously collected experiences without further interaction with the environment.

Qualitative Possibilistic Mixed-Observable MDPs

no code implementations26 Sep 2013 Nicolas Drougard, Florent Teichteil-Konigsbuch, Jean-Loup Farges, Didier Dubois

Possibilistic and qualitative POMDPs (pi-POMDPs) are counterparts of POMDPs used to model situations where the agent's initial belief or observation probabilities are imprecise due to lack of past experiences or insufficient data collection.

Cannot find the paper you are looking for? You can Submit a new open access paper.