1 code implementation • 9 Feb 2024 • Simone Parisi, Montaser Mohammedalamen, Alireza Kazemipour, Matthew E. Taylor, Michael Bowling
In this paper, we formalize a novel but general RL framework - Monitored MDPs - where the agent cannot always observe rewards.
no code implementations • 7 Mar 2022 • Simone Parisi, Aravind Rajeswaran, Senthil Purushwalkam, Abhinav Gupta
In this context, we revisit and study the role of pre-trained visual representations for control, and in particular representations trained on large-scale computer vision datasets.
1 code implementation • NeurIPS 2021 • Simone Parisi, Victoria Dean, Deepak Pathak, Abhinav Gupta
In this setup, the agent first learns to explore across many environments without any extrinsic goal in a task-agnostic manner.
1 code implementation • 1 Jan 2020 • Simone Parisi, Davide Tateo, Maximilian Hensel, Carlo D'Eramo, Jan Peters, Joni Pajarinen
Empirical results on classic and novel benchmarks show that the proposed approach outperforms existing methods in environments with sparse rewards, especially in the presence of rewards that create suboptimal modes of the objective function.
1 code implementation • 19 Dec 2018 • Simone Parisi, Voot Tangkaratt, Jan Peters, Mohammad Emtiyaz Khan
Actor-critic methods can achieve incredible performance on difficult reinforcement learning problems, but they are also prone to instability.
no code implementations • 10 Nov 2016 • Voot Tangkaratt, Herke van Hoof, Simone Parisi, Gerhard Neumann, Jan Peters, Masashi Sugiyama
A naive application of unsupervised dimensionality reduction methods to the context variables, such as principal component analysis, is insufficient as task-relevant input may be ignored.
no code implementations • 13 Jun 2014 • Matteo Pirotta, Simone Parisi, Marcello Restelli
The paper is about learning a continuous approximation of the Pareto frontier in Multi-Objective Markov Decision Problems (MOMDPs).
Multi-Objective Reinforcement Learning reinforcement-learning