Search Results for author: Anders Jonsson

Found 33 papers, 11 papers with code

Planning with a Learned Policy Basis to Optimally Solve Complex Tasks

no code implementations22 Mar 2024 Guillermo Infante, David Kuric, Anders Jonsson, Vicenç Gómez, Herke van Hoof

Conventional reinforcement learning (RL) methods can successfully solve a wide range of sequential decision problems.

Reinforcement Learning (RL)

Asymmetric Norms to Approximate the Minimum Action Distance

1 code implementation16 Dec 2023 Lorenzo Steccanella, Anders Jonsson

Our results show that our asymmetric norm parametrization performs comparably to symmetric norms in symmetric environments and surpasses symmetric norms in asymmetric environments.

Generalized Planning as Heuristic Search: A new planning search-space that leverages pointers over objects

2 code implementations26 Jan 2023 Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson

First, the paper introduces a new pointer-based solution space for GP that is independent of the number of classical planning instances in a GP problem and the size of those instances (i. e. the number of objects, state variables and their domain sizes).

Hierarchies of Reward Machines

1 code implementation31 May 2022 Daniel Furelos-Blanco, Mark Law, Anders Jonsson, Krysia Broda, Alessandra Russo

Reward machines (RMs) are a recent formalism for representing the reward function of a reinforcement learning task through a finite-state machine whose edges encode subgoals of the task using high-level events.

Computing Programs for Generalized Planning as Heuristic Search

no code implementations12 May 2022 Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson

This paper adapts the planning as heuristic search paradigm to the particularities of GP, and presents the first native heuristic search approach to GP.

Scaling-up Generalized Planning as Heuristic Search with Landmarks

no code implementations10 May 2022 Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson, Laura Sebastiá

In this paper we define a landmark counting heuristic for GP (that considers sub-goal information that is not explicitly represented in the planning instances), and a novel heuristic search algorithm for GP (that we call PGP) and that progressively processes subsets of the planning instances of a GP problem.

Hierarchical Representation Learning for Markov Decision Processes

no code implementations3 Jun 2021 Lorenzo Steccanella, Simone Totaro, Anders Jonsson

In this paper we present a novel method for learning hierarchical representations of Markov decision processes.

Representation Learning

Generalized Planning as Heuristic Search

no code implementations26 Mar 2021 Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson

First, the paper defines a novel GP solution space that is independent of the number of planning instances in a GP problem, and the size of these instances.

Hierarchical Width-Based Planning and Learning

1 code implementation15 Jan 2021 Miquel Junyent, Vicenç Gómez, Anders Jonsson

In classical planning, we show how IW(1) at two levels of abstraction can solve problems of width 2.

Atari Games

Hierarchical reinforcement learning for efficient exploration and transfer

no code implementations12 Nov 2020 Lorenzo Steccanella, Simone Totaro, Damien Allonsius, Anders Jonsson

Sparse-reward domains are challenging for reinforcement learning algorithms since significant exploration is needed before encountering reward for the first time.

Efficient Exploration Hierarchical Reinforcement Learning +2

Improved Exploration in Factored Average-Reward MDPs

no code implementations9 Sep 2020 Mohammad Sadegh Talebi, Anders Jonsson, Odalric-Ambrym Maillard

We consider a regret minimization task under the average-reward criterion in an unknown Factored Markov Decision Process (FMDP).

Adaptive Reward-Free Exploration

no code implementations11 Jun 2020 Emilie Kaufmann, Pierre Ménard, Omar Darwiche Domingues, Anders Jonsson, Edouard Leurent, Michal Valko

Reward-free exploration is a reinforcement learning setting studied by Jin et al. (2020), who address it by running several algorithms with regret guarantees in parallel.

Planning in Markov Decision Processes with Gap-Dependent Sample Complexity

no code implementations NeurIPS 2020 Anders Jonsson, Emilie Kaufmann, Pierre Ménard, Omar Darwiche Domingues, Edouard Leurent, Michal Valko

We propose MDP-GapE, a new trajectory-based Monte-Carlo Tree Search algorithm for planning in a Markov Decision Process in which transitions have a finite support.

Usage of Network Simulators in Machine-Learning-Assisted 5G/6G Networks

2 code implementations17 May 2020 Francesc Wilhelmi, Marc Carrascosa, Cristina Cano, Anders Jonsson, Vishnu Ram, Boris Bellalta

Without any doubt, Machine Learning (ML) will be an important driver of future communications due to its foreseen performance when applied to complex problems.

BIG-bench Machine Learning

Generalized Planning with Positive and Negative Examples

1 code implementation21 Nov 2019 Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson

Generalized planning aims at computing an algorithm-like structure (generalized plan) that solves a set of multiple planning instances.

Hierarchical Finite State Controllers for Generalized Planning

no code implementations7 Nov 2019 Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson

We show that hierarchical FSCs can represent generalized plans more compactly than individual FSCs.

Generalized Planning With Procedural Domain Control Knowledge

no code implementations11 Oct 2019 Javier Segovia-Aguas, Sergio Jiménez, Anders Jonsson

Generalized planning is the task of generating a single solution that is valid for a set of planning problems.

valid

Solving Multiagent Planning Problems with Concurrent Conditional Effects

1 code implementation19 Jun 2019 Daniel Furelos-Blanco, Anders Jonsson

Empirically, we show that our compilation can solve challenging multiagent planning problems that require concurrent actions.

Deep Policies for Width-Based Planning in Pixel Domains

1 code implementation12 Apr 2019 Miquel Junyent, Anders Jonsson, Vicenç Gómez

Surprisingly, we observe that the representation learned by the neural network can be used as a feature space for the width-based planner without degrading its performance, thus removing the requirement of pre-defined features for the planner.

Atari Games

Improving width-based planning with compact policies

no code implementations15 Jun 2018 Miquel Junyent, Anders Jonsson, Vicenç Gómez

The planning step hinges on the Iterated-Width (IW) planner, a state of the art planner that makes explicit use of the state representation to perform structured exploration.

Atari Games reinforcement-learning +1

Assessing the impact of machine intelligence on human behaviour: an interdisciplinary endeavour

no code implementations7 Jun 2018 Emilia Gómez, Carlos Castillo, Vicky Charisi, Verónica Dahl, Gustavo Deco, Blagoj Delipetrev, Nicole Dewandre, Miguel Ángel González-Ballester, Fabien Gouyon, José Hernández-Orallo, Perfecto Herrera, Anders Jonsson, Ansgar Koene, Martha Larson, Ramón López de Mántaras, Bertin Martens, Marius Miron, Rubén Moreno-Bote, Nuria Oliver, Antonio Puertas Gallardo, Heike Schweitzer, Nuria Sebastian, Xavier Serra, Joan Serrà, Songül Tolan, Karina Vold

The workshop gathered an interdisciplinary group of experts to establish the state of the art research in the field and a list of future research challenges to be addressed on the topic of human and machine intelligence, algorithm's potential impact on human cognitive capabilities and decision making, and evaluation and regulation needs.

Decision Making

Implications of Decentralized Q-learning Resource Allocation in Wireless Networks

1 code implementation30 May 2017 Francesc Wilhelmi, Boris Bellalta, Cristina Cano, Anders Jonsson

Reinforcement Learning is gaining attention by the wireless networking community due to its potential to learn good-performing configurations only from the observed results.

Q-Learning

A unified view of entropy-regularized Markov decision processes

no code implementations22 May 2017 Gergely Neu, Anders Jonsson, Vicenç Gómez

We propose a general framework for entropy-regularized average-reward reinforcement learning in Markov decision processes (MDPs).

Policy Gradient Methods reinforcement-learning +1

Hierarchical Linearly-Solvable Markov Decision Problems

no code implementations10 Mar 2016 Anders Jonsson, Vicenç Gómez

We present a hierarchical reinforcement learning framework that formulates each task in the hierarchy as a special type of Markov decision process for which the Bellman equation is linear and has analytical solution.

Hierarchical Reinforcement Learning reinforcement-learning +1

Planning over Chain Causal Graphs for Variables with Domains of Size 5 Is NP-Hard

no code implementations15 Jan 2014 Omer Giménez, Anders Jonsson

Recently, considerable focus has been given to the problem of determining the boundary between tractable and intractable planning problems.

The Role of Macros in Tractable Planning

no code implementations15 Jan 2014 Anders Jonsson

We generalize the inverted tree reducible class in several ways and describe modifications of the algorithm to deal with these new classes.

Cannot find the paper you are looking for? You can Submit a new open access paper.