Search Results for author: Carlo D'Eramo

Found 20 papers, 7 papers with code

Generalized Mean Estimation in Monte-Carlo Tree Search

no code implementations1 Nov 2019 Tuan Dam, Pascal Klink, Carlo D'Eramo, Jan Peters, Joni Pajarinen

Finally, we empirically demonstrate the effectiveness of our method in well-known MDP and POMDP benchmarks, showing significant improvement in performance and convergence speed w. r. t.

Long-Term Visitation Value for Deep Exploration in Sparse Reward Reinforcement Learning

1 code implementation1 Jan 2020 Simone Parisi, Davide Tateo, Maximilian Hensel, Carlo D'Eramo, Jan Peters, Joni Pajarinen

Empirical results on classic and novel benchmarks show that the proposed approach outperforms existing methods in environments with sparse rewards, especially in the presence of rewards that create suboptimal modes of the objective function.

Benchmarking reinforcement-learning +1

MushroomRL: Simplifying Reinforcement Learning Research

2 code implementations4 Jan 2020 Carlo D'Eramo, Davide Tateo, Andrea Bonarini, Marcello Restelli, Jan Peters

MushroomRL is an open-source Python library developed to simplify the process of implementing and running Reinforcement Learning (RL) experiments.

reinforcement-learning Reinforcement Learning (RL)

Deep Reinforcement Learning with Weighted Q-Learning

no code implementations20 Mar 2020 Andrea Cini, Carlo D'Eramo, Jan Peters, Cesare Alippi

In this regard, Weighted Q-Learning (WQL) effectively reduces bias and shows remarkable results in stochastic environments.

Gaussian Processes Q-Learning +3

Self-Paced Deep Reinforcement Learning

1 code implementation NeurIPS 2020 Pascal Klink, Carlo D'Eramo, Jan Peters, Joni Pajarinen

Curriculum reinforcement learning (CRL) improves the learning speed and stability of an agent by exposing it to a tailored series of tasks throughout learning.

Open-Ended Question Answering reinforcement-learning +1

Convex Regularization in Monte-Carlo Tree Search

no code implementations1 Jul 2020 Tuan Dam, Carlo D'Eramo, Jan Peters, Joni Pajarinen

Monte-Carlo planning and Reinforcement Learning (RL) are essential to sequential decision making.

Atari Games Decision Making +1

A Probabilistic Interpretation of Self-Paced Learning with Applications to Reinforcement Learning

1 code implementation25 Feb 2021 Pascal Klink, Hany Abdulsamad, Boris Belousov, Carlo D'Eramo, Jan Peters, Joni Pajarinen

Across machine learning, the use of curricula has shown strong empirical potential to improve learning from data by avoiding local optima of training objectives.

reinforcement-learning Reinforcement Learning (RL)

Model Predictive Actor-Critic: Accelerating Robot Skill Acquisition with Deep Reinforcement Learning

1 code implementation25 Mar 2021 Andrew S. Morgan, Daljeet Nandha, Georgia Chalvatzaki, Carlo D'Eramo, Aaron M. Dollar, Jan Peters

Substantial advancements to model-based reinforcement learning algorithms have been impeded by the model-bias induced by the collected data, which generally hurts performance.

Model-based Reinforcement Learning Model Predictive Control +2

Boosted Curriculum Reinforcement Learning

no code implementations ICLR 2022 Pascal Klink, Carlo D'Eramo, Jan Peters, Joni Pajarinen

This approach, which we refer to as boosted curriculum reinforcement learning (BCRL), has the benefit of naturally increasing the representativeness of the functional space by adding a new residual each time a new task is presented.

reinforcement-learning Reinforcement Learning (RL)

A Unified Perspective on Value Backup and Exploration in Monte-Carlo Tree Search

no code implementations11 Feb 2022 Tuan Dam, Carlo D'Eramo, Jan Peters, Joni Pajarinen

In this work, we propose two methods for improving the convergence rate and exploration based on a newly introduced backup operator and entropy regularization.

Atari Games Decision Making +2

Monte-Carlo tree search with uncertainty propagation via optimal transport

no code implementations19 Sep 2023 Tuan Dam, Pascal Stenger, Lukas Schneider, Joni Pajarinen, Carlo D'Eramo, Odalric-Ambrym Maillard

We introduce a novel backup operator that computes value nodes as the Wasserstein barycenter of their action-value children nodes; thus, propagating the uncertainty of the estimate across the tree to the root node.

Thompson Sampling

On the Benefit of Optimal Transport for Curriculum Reinforcement Learning

no code implementations25 Sep 2023 Pascal Klink, Carlo D'Eramo, Jan Peters, Joni Pajarinen

In this work, we focus on framing curricula as interpolations between task distributions, which has previously been shown to be a viable approach to CRL.

reinforcement-learning

Domain Randomization via Entropy Maximization

no code implementations3 Nov 2023 Gabriele Tiboni, Pascal Klink, Jan Peters, Tatiana Tommasi, Carlo D'Eramo, Georgia Chalvatzaki

Varying dynamics parameters in simulation is a popular Domain Randomization (DR) approach for overcoming the reality gap in Reinforcement Learning (RL).

Reinforcement Learning (RL)

Robust Adversarial Reinforcement Learning via Bounded Rationality Curricula

no code implementations3 Nov 2023 Aryaman Reddi, Maximilian Tölle, Jan Peters, Georgia Chalvatzaki, Carlo D'Eramo

To this end, Robust Adversarial Reinforcement Learning (RARL) trains a protagonist against destabilizing forces exercised by an adversary in a competitive zero-sum Markov game, whose optimal solution, i. e., rational strategy, corresponds to a Nash equilibrium.

reinforcement-learning Reinforcement Learning (RL)

Multi-Task Reinforcement Learning with Mixture of Orthogonal Experts

no code implementations19 Nov 2023 Ahmed Hendawy, Jan Peters, Carlo D'Eramo

Multi-Task Reinforcement Learning (MTRL) tackles the long-standing problem of endowing agents with skills that generalize across a variety of problems.

reinforcement-learning Representation Learning

Contact Energy Based Hindsight Experience Prioritization

no code implementations5 Dec 2023 Erdi Sayar, Zhenshan Bing, Carlo D'Eramo, Ozgur S. Oguz, Alois Knoll

Multi-goal robot manipulation tasks with sparse rewards are difficult for reinforcement learning (RL) algorithms due to the inefficiency in collecting successful experiences.

Reinforcement Learning (RL) Robot Manipulation

Parameterized Projected Bellman Operator

1 code implementation20 Dec 2023 Théo Vincent, Alberto Maria Metelli, Boris Belousov, Jan Peters, Marcello Restelli, Carlo D'Eramo

We formulate an optimization problem to learn PBO for generic sequential decision-making problems, and we theoretically analyze its properties in two representative classes of RL problems.

Decision Making Reinforcement Learning (RL)

Sharing Knowledge in Multi-Task Deep Reinforcement Learning

1 code implementation ICLR 2020 Carlo D'Eramo, Davide Tateo, Andrea Bonarini, Marcello Restelli, Jan Peters

We study the benefit of sharing representations among tasks to enable the effective use of deep neural networks in Multi-Task Reinforcement Learning.

reinforcement-learning

Iterated $Q$-Network: Beyond the One-Step Bellman Operator

no code implementations4 Mar 2024 Théo Vincent, Daniel Palenicek, Boris Belousov, Jan Peters, Carlo D'Eramo

Value-based Reinforcement Learning (RL) methods rely on the application of the Bellman operator, which needs to be approximated from samples.

Atari Games Continuous Control +1

Cannot find the paper you are looking for? You can Submit a new open access paper.