no code implementations • 4 Sep 2024 • Ahana Deb, Roberto Cipollone, Anders Jonsson, Alessandro Ronca, Mohammad Sadegh Talebi
In this paper, we show that it is possible to overcome two strong limitations of previous offline RL algorithms for RDPs, notably RegORL.
no code implementations • 9 Jul 2024 • Guillermo Infante, Anders Jonsson, Vicenç Gómez
We introduce a novel approach to hierarchical reinforcement learning for Linearly-solvable Markov Decision Processes (LMDPs) in the infinite-horizon average-reward setting.
no code implementations • 6 Jun 2024 • Sergio Calo, Anders Jonsson, Gergely Neu, Ludovic Schwartz, Javier Segovia
We propose a new framework for formulating optimal transport distances between Markov chains.
no code implementations • 22 Mar 2024 • Guillermo Infante, David Kuric, Anders Jonsson, Vicenç Gómez, Herke van Hoof
Conventional reinforcement learning (RL) methods can successfully solve a wide range of sequential decision problems.
1 code implementation • 16 Dec 2023 • Lorenzo Steccanella, Anders Jonsson
Our results show that our asymmetric norm parametrization performs comparably to symmetric norms in symmetric environments and surpasses symmetric norms in asymmetric environments.
2 code implementations • 26 Jan 2023 • Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson
First, the paper introduces a new pointer-based solution space for GP that is independent of the number of classical planning instances in a GP problem and the size of those instances (i. e. the number of objects, state variables and their domain sizes).
1 code implementation • 31 May 2022 • Daniel Furelos-Blanco, Mark Law, Anders Jonsson, Krysia Broda, Alessandra Russo
Reward machines (RMs) are a recent formalism for representing the reward function of a reinforcement learning task through a finite-state machine whose edges encode subgoals of the task using high-level events.
no code implementations • 12 May 2022 • Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson
This paper adapts the planning as heuristic search paradigm to the particularities of GP, and presents the first native heuristic search approach to GP.
no code implementations • 10 May 2022 • Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson, Laura Sebastiá
In this paper we define a landmark counting heuristic for GP (that considers sub-goal information that is not explicitly represented in the planning instances), and a novel heuristic search algorithm for GP (that we call PGP) and that progressively processes subsets of the planning instances of a GP problem.
no code implementations • 4 May 2022 • Lorenzo Steccanella, Anders Jonsson
This paper presents a novel state representation for reward-free Markov decision processes.
1 code implementation • 29 Jun 2021 • Guillermo Infante, Anders Jonsson, Vicenç Gómez
In this work we present a novel approach to hierarchical reinforcement learning for linearly-solvable Markov decision processes.
Hierarchical Reinforcement Learning reinforcement-learning +1
no code implementations • 3 Jun 2021 • Lorenzo Steccanella, Simone Totaro, Anders Jonsson
In this paper we present a novel method for learning hierarchical representations of Markov decision processes.
no code implementations • 26 Mar 2021 • Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson
First, the paper defines a novel GP solution space that is independent of the number of planning instances in a GP problem, and the size of these instances.
1 code implementation • 15 Jan 2021 • Miquel Junyent, Vicenç Gómez, Anders Jonsson
In classical planning, we show how IW(1) at two levels of abstraction can solve problems of width 2.
no code implementations • 12 Nov 2020 • Lorenzo Steccanella, Simone Totaro, Damien Allonsius, Anders Jonsson
Sparse-reward domains are challenging for reinforcement learning algorithms since significant exploration is needed before encountering reward for the first time.
Efficient Exploration Hierarchical Reinforcement Learning +2
no code implementations • 9 Sep 2020 • Mohammad Sadegh Talebi, Anders Jonsson, Odalric-Ambrym Maillard
We consider a regret minimization task under the average-reward criterion in an unknown Factored Markov Decision Process (FMDP).
no code implementations • 8 Sep 2020 • Daniel Furelos-Blanco, Mark Law, Anders Jonsson, Krysia Broda, Alessandra Russo
In this paper we present ISA, an approach for learning and exploiting subgoals in episodic reinforcement learning (RL) tasks.
no code implementations • 27 Jul 2020 • Pierre Ménard, Omar Darwiche Domingues, Anders Jonsson, Emilie Kaufmann, Edouard Leurent, Michal Valko
Realistic environments often provide agents with very limited feedback.
no code implementations • ICML Workshop LifelongML 2020 • Lorenzo Steccanella, Simone Totaro, Damien Allonsius, Anders Jonsson
Sparse-reward domains are challenging for reinforcement learning algorithms since significant exploration is needed before encountering reward for the first time.
Hierarchical Reinforcement Learning reinforcement-learning +1
no code implementations • 11 Jun 2020 • Emilie Kaufmann, Pierre Ménard, Omar Darwiche Domingues, Anders Jonsson, Edouard Leurent, Michal Valko
Reward-free exploration is a reinforcement learning setting studied by Jin et al. (2020), who address it by running several algorithms with regret guarantees in parallel.
no code implementations • NeurIPS 2020 • Anders Jonsson, Emilie Kaufmann, Pierre Ménard, Omar Darwiche Domingues, Edouard Leurent, Michal Valko
We propose MDP-GapE, a new trajectory-based Monte-Carlo Tree Search algorithm for planning in a Markov Decision Process in which transitions have a finite support.
2 code implementations • 17 May 2020 • Francesc Wilhelmi, Marc Carrascosa, Cristina Cano, Anders Jonsson, Vishnu Ram, Boris Bellalta
Without any doubt, Machine Learning (ML) will be an important driver of future communications due to its foreseen performance when applied to complex problems.
1 code implementation • 16 May 2020 • Simone Totaro, Ioannis Boukas, Anders Jonsson, Bertrand Cornélusse
We propose a novel model based reinforcement learning algorithm that is able to address both types of changes.
Model-based Reinforcement Learning reinforcement-learning +1
no code implementations • 29 Nov 2019 • Daniel Furelos-Blanco, Mark Law, Alessandra Russo, Krysia Broda, Anders Jonsson
In this work we present ISA, a novel approach for learning and exploiting subgoals in reinforcement learning (RL).
1 code implementation • 21 Nov 2019 • Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson
Generalized planning aims at computing an algorithm-like structure (generalized plan) that solves a set of multiple planning instances.
no code implementations • 7 Nov 2019 • Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson
We show that hierarchical FSCs can represent generalized plans more compactly than individual FSCs.
no code implementations • 11 Oct 2019 • Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson
Generalized planning is the task of generating a single solution that is valid for a set of planning problems.
1 code implementation • 19 Jun 2019 • Daniel Furelos-Blanco, Anders Jonsson
Empirically, we show that our compilation can solve challenging multiagent planning problems that require concurrent actions.
1 code implementation • 12 Apr 2019 • Miquel Junyent, Anders Jonsson, Vicenç Gómez
Surprisingly, we observe that the representation learned by the neural network can be used as a feature space for the width-based planner without degrading its performance, thus removing the requirement of pre-defined features for the planner.
no code implementations • 15 Jun 2018 • Miquel Junyent, Anders Jonsson, Vicenç Gómez
The planning step hinges on the Iterated-Width (IW) planner, a state of the art planner that makes explicit use of the state representation to perform structured exploration.
no code implementations • 7 Jun 2018 • Emilia Gómez, Carlos Castillo, Vicky Charisi, Verónica Dahl, Gustavo Deco, Blagoj Delipetrev, Nicole Dewandre, Miguel Ángel González-Ballester, Fabien Gouyon, José Hernández-Orallo, Perfecto Herrera, Anders Jonsson, Ansgar Koene, Martha Larson, Ramón López de Mántaras, Bertin Martens, Marius Miron, Rubén Moreno-Bote, Nuria Oliver, Antonio Puertas Gallardo, Heike Schweitzer, Nuria Sebastian, Xavier Serra, Joan Serrà, Songül Tolan, Karina Vold
The workshop gathered an interdisciplinary group of experts to establish the state of the art research in the field and a list of future research challenges to be addressed on the topic of human and machine intelligence, algorithm's potential impact on human cognitive capabilities and decision making, and evaluation and regulation needs.
1 code implementation • 30 May 2017 • Francesc Wilhelmi, Boris Bellalta, Cristina Cano, Anders Jonsson
Reinforcement Learning is gaining attention by the wireless networking community due to its potential to learn good-performing configurations only from the observed results.
no code implementations • 22 May 2017 • Gergely Neu, Anders Jonsson, Vicenç Gómez
We propose a general framework for entropy-regularized average-reward reinforcement learning in Markov decision processes (MDPs).
no code implementations • 10 Mar 2016 • Anders Jonsson, Vicenç Gómez
We present a hierarchical reinforcement learning framework that formulates each task in the hierarchy as a special type of Markov decision process for which the Bellman equation is linear and has analytical solution.
Hierarchical Reinforcement Learning reinforcement-learning +1
no code implementations • 15 Jan 2014 • Anders Jonsson
We generalize the inverted tree reducible class in several ways and describe modifications of the algorithm to deal with these new classes.
no code implementations • 15 Jan 2014 • Omer Giménez, Anders Jonsson
Recently, considerable focus has been given to the problem of determining the boundary between tractable and intractable planning problems.