no code implementations • 5 Sep 2024 • Rongzhen Zhao, Vivienne Wang, Juho Kannala, Joni Pajarinen
We propose Organized GDR (OGDR) to organize channels belonging to the same attributes together for correct decomposition from features into attributes.
no code implementations • 20 Aug 2024 • Yi Zhao, Le Chen, Jan Schneider, Quankai Gao, Juho Kannala, Bernhard Schölkopf, Joni Pajarinen, Dieter Büchler
It has been a long-standing research goal to endow robot hands with human-level dexterity.
1 code implementation • 4 Jul 2024 • Changling Li, Zhang-Wei Hong, Pulkit Agrawal, Divyansh Garg, Joni Pajarinen
We specifically explore the KL divergence as the regularizer and obtain a new form of prioritization scheme, the regularized optimal experience replay (ROER).
no code implementations • 1 Jul 2024 • Rongzhen Zhao, Vivienne Wang, Juho Kannala, Joni Pajarinen
Similar to humans perceiving visual scenes as objects, Object-Centric Learning (OCL) can abstract dense images or videos into sparse object-level features.
1 code implementation • 24 Jun 2024 • Vivienne Huiling Wang, Tinghuai Wang, Wenyan Yang, Joni-Kristian Kämäräinen, Joni Pajarinen
In goal-conditioned hierarchical reinforcement learning (HRL), a high-level policy specifies a subgoal for the low-level policy to reach.
1 code implementation • 12 Jun 2024 • Mohammadreza Nakhaei, Aidan Scannell, Joni Pajarinen
In general, these methods assume that the transition dynamics remain the same during both the offline and online phases of training.
no code implementations • 4 Jun 2024 • Aidan Scannell, Kalle Kujanpää, Yi Zhao, Mohammadreza Nakhaei, Arno Solin, Joni Pajarinen
Learning representations for reinforcement learning (RL) has shown much promise for continuous control.
1 code implementation • 16 Mar 2024 • Aidan Scannell, Riccardo Mereu, Paul Chang, Ella Tamir, Joni Pajarinen, Arno Solin
Our parameterization offers: (i) a way to scale function-space methods to large data sets via sparsification, (ii) retention of prior knowledge when access to past data is limited, and (iii) a mechanism to incorporate new data without retraining.
no code implementations • 16 Jan 2024 • Zhiyuan Li, Wenshuai Zhao, Lijun Wu, Joni Pajarinen
Inspired by the concept of correlated equilibrium, we propose to introduce a \textit{strategy modification} to provide a mechanism for agents to correlate their policies.
1 code implementation • 3 Nov 2023 • Wenshuai Zhao, Yi Zhao, Zhiyuan Li, Juho Kannala, Joni Pajarinen
*Relative overgeneralization* (RO) occurs in cooperative multi-agent learning tasks when agents converge towards a suboptimal joint policy due to overfitting to suboptimal behavior of other agents.
1 code implementation • NeurIPS 2023 • Zhang-Wei Hong, Aviral Kumar, Sathwik Karnik, Abhishek Bhandwaldar, Akash Srivastava, Joni Pajarinen, Romain Laroche, Abhishek Gupta, Pulkit Agrawal
We argue this is due to an assumption made by current offline RL algorithms of staying close to the trajectories in the dataset.
no code implementations • 25 Sep 2023 • Pascal Klink, Florian Wolf, Kai Ploeger, Jan Peters, Joni Pajarinen
Reinforcement Learning (RL) allows learning non-trivial robot control laws purely from data.
no code implementations • 25 Sep 2023 • Pascal Klink, Carlo D'Eramo, Jan Peters, Joni Pajarinen
In this work, we focus on framing curricula as interpolations between task distributions, which has previously been shown to be a viable approach to CRL.
no code implementations • 19 Sep 2023 • Tuan Dam, Pascal Stenger, Lukas Schneider, Joni Pajarinen, Carlo D'Eramo, Odalric-Ambrym Maillard
We introduce a novel backup operator that computes value nodes as the Wasserstein barycenter of their action-value children nodes; thus, propagating the uncertainty of the estimate across the tree to the root node.
2 code implementations • 5 Sep 2023 • Aidan Scannell, Riccardo Mereu, Paul Chang, Ella Tamir, Joni Pajarinen, Arno Solin
Deep neural networks (NNs) are known to lack uncertainty estimates and struggle to incorporate new data.
no code implementations • 31 Jul 2023 • Abdolreza Taheri, Robert Pettersson, Pelle Gustafsson, Joni Pajarinen, Reza Ghabcheloo
A sizable part of the fleet of heavy-duty machinery in the construction equipment industry uses the conventional valve-controlled load-sensing hydraulics.
1 code implementation • 15 Jun 2023 • Yi Zhao, Wenshuai Zhao, Rinu Boney, Juho Kannala, Joni Pajarinen
This applies when using pure planning with a dynamics model conditioned on the representation, but, also when utilizing the representation as policy and value function features in model-free RL.
1 code implementation • 17 Feb 2023 • Wenyan Yang, Huiling Wang, Dingding Cai, Joni Pajarinen, Joni-Kristen Kämäräinen
Offline goal-conditioned reinforcement learning (GCRL) can be challenging due to overfitting to the given dataset.
no code implementations • 15 Feb 2023 • Wenyan Yang, Joni Pajarinen, Dinging Cai, Joni Kämäräinen
In goal-conditioned offline reinforcement learning, an agent learns from previously collected data to go to an arbitrary goal.
1 code implementation • 30 Jan 2023 • Kalle Kujanpää, Joni Pajarinen, Alexander Ilin
The ability to plan actions on multiple levels of abstraction enables intelligent agents to solve complex tasks effectively.
1 code implementation • 14 Nov 2022 • Eric Chen, Zhang-Wei Hong, Joni Pajarinen, Pulkit Agrawal
However, on easy exploration tasks, the agent gets distracted by intrinsic rewards and performs unnecessary exploration even when sufficient task (also called extrinsic) reward is available.
2 code implementations • 25 Oct 2022 • Yi Zhao, Rinu Boney, Alexander Ilin, Juho Kannala, Joni Pajarinen
Offline reinforcement learning, by learning from a fixed dataset, makes it possible to learn agent behaviors without interacting with the environment.
1 code implementation • 4 Oct 2022 • Kalle Kujanpää, Amin Babadi, Yi Zhao, Juho Kannala, Alexander Ilin, Joni Pajarinen
To address this problem, we propose Continuous Monte Carlo Graph Search (CMCGS), an extension of MCTS to online planning in environments with continuous state and action spaces.
no code implementations • 21 Sep 2022 • Mikko Lauri, David Hsu, Joni Pajarinen
Noisy sensing, imperfect control, and environment changes are defining characteristics of many real-world robot tasks.
no code implementations • 20 May 2022 • Wenshuai Zhao, Zhiyuan Li, Joni Pajarinen
Inspired by the success of CRL in single-agent settings, a few works have attempted to apply CRL to multi-agent reinforcement learning (MARL) using the number of agents to control task difficulty.
Multi-agent Reinforcement Learning Open-Ended Question Answering +3
1 code implementation • ICLR 2022 • Zhang-Wei Hong, Tao Chen, Yen-Chen Lin, Joni Pajarinen, Pulkit Agrawal
State-of-the-art deep Q-learning methods update Q-values using state transition tuples sampled from the experience replay buffer.
no code implementations • 28 Feb 2022 • Abdolreza Taheri, Joni Pajarinen, Reza Ghabcheloo
The ability of Gaussian processes (GPs) to predict the behavior of dynamical systems as a more sample-efficient alternative to parametric models seems promising for real-world robotics research.
no code implementations • 11 Feb 2022 • Tuan Dam, Carlo D'Eramo, Jan Peters, Joni Pajarinen
In this work, we propose two methods for improving the convergence rate and exploration based on a newly introduced backup operator and entropy regularization.
no code implementations • 24 Jan 2022 • Vivienne Huiling Wang, Joni Pajarinen, Tinghuai Wang, Joni-Kristian Kämäräinen
Hierarchical reinforcement learning (HRL) proposes to solve difficult tasks by performing decision-making and control at successively higher levels of temporal abstraction.
no code implementations • 29 Sep 2021 • Pascal Klink, Haoyi Yang, Jan Peters, Joni Pajarinen
Experiments demonstrate that the resulting introduction of metric structure into the curriculum allows for a well-behaving non-parametric version of SPRL that leads to stable learning performance across tasks.
no code implementations • ICLR 2022 • Pascal Klink, Carlo D'Eramo, Jan Peters, Joni Pajarinen
This approach, which we refer to as boosted curriculum reinforcement learning (BCRL), has the benefit of naturally increasing the representativeness of the functional space by adding a new residual each time a new task is presented.
no code implementations • 22 Apr 2021 • Stephan Weigand, Pascal Klink, Jan Peters, Joni Pajarinen
Due to recent breakthroughs, reinforcement learning (RL) has demonstrated impressive performance in challenging sequential decision-making problems.
no code implementations • 23 Mar 2021 • Wenyan Yang, Nataliya Strokina, Nikolay Serbenyuk, Joni Pajarinen, Reza Ghabcheloo, Juho Vihonen, Mohammad M. Aref, Joni-Kristian Kämäräinen
We have recently proposed two pile loading controllers that learn from human demonstrations: a neural network (NNet) [1] and a random forest (RF) controller [2].
1 code implementation • 25 Feb 2021 • Pascal Klink, Hany Abdulsamad, Boris Belousov, Carlo D'Eramo, Jan Peters, Joni Pajarinen
Across machine learning, the use of curricula has shown strong empirical potential to improve learning from data by avoiding local optima of training objectives.
no code implementations • 4 Sep 2020 • Joni Pajarinen
The policy graph improvement (PGI) algorithm for POMDPs represents the policy as a fixed size policy graph and improves the policy monotonically.
no code implementations • 4 Jul 2020 • Mikko Lauri, Joni Pajarinen, Jan Peters, Simone Frintrop
We consider the problem of creating a 3D model using depth images captured by a team of multiple robots.
no code implementations • 1 Jul 2020 • Tuan Dam, Carlo D'Eramo, Jan Peters, Joni Pajarinen
Monte-Carlo planning and Reinforcement Learning (RL) are essential to sequential decision making.
no code implementations • 27 Apr 2020 • Lauri Alho, Adrian Burian, Janne Helenius, Joni Pajarinen
Identifying mobile network problems in 4G cells is more challenging when the complexity of the network increases, and privacy concerns limit the information content of the data.
1 code implementation • NeurIPS 2020 • Pascal Klink, Carlo D'Eramo, Jan Peters, Joni Pajarinen
Curriculum reinforcement learning (CRL) improves the learning speed and stability of an agent by exposing it to a tailored series of tasks throughout learning.
no code implementations • 8 Mar 2020 • Melvin Laux, Oleg Arenz, Jan Peters, Joni Pajarinen
The ARL framework utilizes an adversary, which is trained to steer the original agent, the protagonist, to challenging states.
1 code implementation • 1 Jan 2020 • Simone Parisi, Davide Tateo, Maximilian Hensel, Carlo D'Eramo, Jan Peters, Joni Pajarinen
Empirical results on classic and novel benchmarks show that the proposed approach outperforms existing methods in environments with sparse rewards, especially in the presence of rewards that create suboptimal modes of the objective function.
no code implementations • 1 Nov 2019 • Tuan Dam, Pascal Klink, Carlo D'Eramo, Jan Peters, Joni Pajarinen
Finally, we empirically demonstrate the effectiveness of our method in well-known MDP and POMDP benchmarks, showing significant improvement in performance and convergence speed w. r. t.
no code implementations • 15 Aug 2019 • Zhang-Wei Hong, Joni Pajarinen, Jan Peters
Model-based Reinforcement Learning (MBRL) allows data-efficient learning which is required in real world applications such as robotics.
1 code implementation • 26 Feb 2019 • Mikko Lauri, Joni Pajarinen, Jan Peters
Decentralized policies for information gathering are required when multiple autonomous agents are deployed to collect data about a phenomenon of interest without the ability to communicate.
no code implementations • 7 Feb 2019 • Joni Pajarinen, Hong Linh Thai, Riad Akrour, Jan Peters, Gerhard Neumann
Trust-region methods have yielded state-of-the-art results in policy search.
no code implementations • 16 Nov 2018 • Takayuki Osa, Joni Pajarinen, Gerhard Neumann, J. Andrew Bagnell, Pieter Abbeel, Jan Peters
This process of learning from demonstrations, and the study of algorithms to do so, is called imitation learning.