1 code implementation • 15 Jun 2023 • Yi Zhao, Wenshuai Zhao, Rinu Boney, Juho Kannala, Joni Pajarinen
This applies when using pure planning with a dynamics model conditioned on the representation, but, also when utilizing the representation as policy and value function features in model-free RL.
2 code implementations • 25 Oct 2022 • Yi Zhao, Rinu Boney, Alexander Ilin, Juho Kannala, Joni Pajarinen
Offline reinforcement learning, by learning from a fixed dataset, makes it possible to learn agent behaviors without interacting with the environment.
2 code implementations • 15 Jun 2021 • Rinu Boney, Alexander Ilin, Juho Kannala
In many control problems that include vision, optimal controls can be inferred from the location of the objects in the scene.
1 code implementation • 22 Dec 2020 • Rinu Boney, Alexander Ilin, Juho Kannala, Jarno Seppänen
We experimentally show that planning with naive Monte Carlo tree search does not perform very well in large combinatorial action spaces.
1 code implementation • 5 Nov 2020 • Rinu Boney, Jussi Sainio, Mikko Kaivola, Arno Solin, Juho Kannala
We validate the platform with reinforcement learning experiments and provide baseline results on a set of benchmark tasks.
no code implementations • 12 Oct 2019 • Rinu Boney, Juho Kannala, Alexander Ilin
Model-based reinforcement learning could enable sample-efficient learning by quickly acquiring rich knowledge about the world and using it to improve behaviour without additional data.
no code implementations • NeurIPS 2019 • Rinu Boney, Norman Di Palo, Mathias Berglund, Alexander Ilin, Juho Kannala, Antti Rasmus, Harri Valpola
Trajectory optimization using a learned model of the environment is one of the core elements of model-based reinforcement learning.
no code implementations • 29 Nov 2017 • Rinu Boney, Alexander Ilin
We consider the problem of semi-supervised few-shot classification where a classifier needs to adapt to new tasks using a few labeled examples and (potentially many) unlabeled examples.
no code implementations • NeurIPS 2017 • Isabeau Prémont-Schwarz, Alexander Ilin, Tele Hotloo Hao, Antti Rasmus, Rinu Boney, Harri Valpola
We propose a recurrent extension of the Ladder networks whose structure is motivated by the inference required in hierarchical latent variable models.