no code implementations • 19 Feb 2025 • Charly Pecqueux-Guézénec, Stéphane Doncieux, Nicolas Perrin-Gilbert
To address the need for safer learning, we propose a method that enables agents to learn goal-conditioned behaviors that explore without the risk of making harmful mistakes.
1 code implementation • 24 Apr 2024 • Nicolas Perrin-Gilbert
This paper presents AFU, an off-policy deep RL algorithm addressing in a new way the challenging "max-Q problem" in Q-learning for continuous action spaces, with a solution based on regression and conditional gradient scaling.
no code implementations • 2 Mar 2024 • Jason Z. Kim, Nicolas Perrin-Gilbert, Erkan Narmanli, Paul Klein, Christopher R. Myers, Itai Cohen, Joshua J. Waterfall, James P. Sethna
Natural systems with emergent behaviors often organize along low-dimensional subsets of high-dimensional spaces.
no code implementations • 14 Feb 2024 • Alexandre Chenu, Olivier Serris, Olivier Sigaud, Nicolas Perrin-Gilbert
Demonstrations are commonly used to speed up the learning process of Deep Reinforcement Learning algorithms.
no code implementations • 1 Nov 2023 • Olivier Sigaud, Gianluca Baldassarre, Cedric Colas, Stephane Doncieux, Richard Duro, Pierre-Yves Oudeyer, Nicolas Perrin-Gilbert, Vieri Giuliano Santucci
A lot of recent machine learning research papers have ``open-ended learning'' in their title.
no code implementations • 13 Jul 2023 • Emily Clement, Nicolas Perrin-Gilbert, Philipp Schlehuber-Caissier
In this paper we present a layered approach for multi-agent control problem, decomposed into three stages, each building upon the results of the previous one.
no code implementations • 27 Mar 2023 • Valentin Macé, Raphaël Boige, Felix Chalumeau, Thomas Pierrot, Guillaume Richard, Nicolas Perrin-Gilbert
In the context of neuroevolution, Quality-Diversity algorithms have proven effective in generating repertoires of diverse and efficient policies by relying on the definition of a behavior space.
no code implementations • 24 Nov 2022 • Felix Chalumeau, Thomas Pierrot, Valentin Macé, Arthur Flajolet, Karim Beguir, Antoine Cully, Nicolas Perrin-Gilbert
Exploration is at the heart of several domains trying to solve control problems such as Reinforcement Learning and QD methods are promising candidates to overcome the challenges associated.
1 code implementation • 9 Nov 2022 • Alexandre Chenu, Olivier Serris, Olivier Sigaud, Nicolas Perrin-Gilbert
This sequential goal-reaching approach raises a problem of compatibility between successive goals: we need to ensure that the state resulting from reaching a goal is compatible with the achievement of the following goals.
1 code implementation • 15 Apr 2022 • Alexandre Chenu, Nicolas Perrin-Gilbert, Olivier Sigaud
In such context, Imitation Learning (IL) can be a powerful approach to bootstrap the learning process.
1 code implementation • 28 Sep 2021 • Astrid Merckling, Nicolas Perrin-Gilbert, Alex Coninx, Stéphane Doncieux
Our experimental results show that the approach leads to efficient exploration in challenging environments with image observations, and to state representations that significantly accelerate learning in RL tasks.
no code implementations • 10 Apr 2021 • Alexandre Chenu, Nicolas Perrin-Gilbert, Stéphane Doncieux, Olivier Sigaud
In particular, we show empirically that, if the mapping is smooth enough, i. e. if two close policies in the parameter space lead to similar outcomes, then diversity algorithms tend to inherit exploration properties of MP algorithms.
1 code implementation • NeurIPS 2021 • Thomas Pierrot, Valentin Macé, Félix Chalumeau, Arthur Flajolet, Geoffrey Cideron, Karim Beguir, Antoine Cully, Olivier Sigaud, Nicolas Perrin-Gilbert
This paper proposes a novel algorithm, QDPG, which combines the strength of Policy Gradient algorithms and Quality Diversity approaches to produce a collection of diverse and high-performing neural policies in continuous control environments.
no code implementations • 15 Sep 2019 • Astrid Merckling, Alexandre Coninx, Loic Cressot, Stéphane Doncieux, Nicolas Perrin-Gilbert
Indeed, a compact representation of such a state is beneficial to help robots grasp onto their environment for interacting.