no code implementations • 16 Jul 2024 • Aske Plaat, Annie Wong, Suzan Verberne, Joost Broekens, Niki van Stein, Thomas Back
The field started with the question whether LLMs can solve grade school math word problems.
no code implementations • 4 Sep 2023 • Joost Broekens, Bernhard Hilpert, Suzan Verberne, Kim Baraka, Patrick Gebhard, Aske Plaat
Large language models, in particular generative pre-trained transformers (GPTs), show impressive results on a wide variety of language-related tasks.
no code implementations • 27 Jan 2023 • Bernd Dudzik, Joost Broekens
In this article, we highlight the important influence of the temporal distance between a stimulus event and the moment when its experience is reported on the provided information's validity.
no code implementations • 27 Aug 2020 • Bernd Dudzik, Joost Broekens, Mark Neerincx, Hayley Hung
A key challenge in the accurate prediction of viewers' emotional responses to video stimuli in real-world applications is accounting for person- and situation-specific variation.
no code implementations • 30 Jun 2020 • Thomas M. Moerland, Joost Broekens, Aske Plaat, Catholijn M. Jonker
Two key approaches to this problem are reinforcement learning (RL) and planning.
no code implementations • 26 Jun 2020 • Thomas M. Moerland, Joost Broekens, Aske Plaat, Catholijn M. Jonker
Therefore, this paper presents a unifying algorithmic framework for reinforcement learning and planning (FRAP), which identifies underlying dimensions on which MDP planning and learning algorithms have to decide.
1 code implementation • 19 May 2020 • Thomas M. Moerland, Joost Broekens, Aske Plaat, Catholijn M. Jonker
Monte Carlo Tree Search (MCTS) efficiently balances exploration and exploitation in tree search based on count-derived uncertainty.
1 code implementation • 15 May 2020 • Thomas M. Moerland, Anna Deichler, Simone Baldi, Joost Broekens, Catholijn M. Jonker
Planning and reinforcement learning are two key approaches to sequential decision making.
no code implementations • 24 Jul 2018 • Joost Broekens
Emotions are intimately tied to motivation and the adaptation of behavior, and many animal species show evidence of emotions in their behavior.
1 code implementation • 11 Jun 2018 • Thomas M. Moerland, Joost Broekens, Catholijn M. Jonker
This paper studies the potential of the return distribution for exploration in deterministic reinforcement learning (RL) environments.
2 code implementations • 24 May 2018 • Thomas M. Moerland, Joost Broekens, Aske Plaat, Catholijn M. Jonker
A core novelty of Alpha Zero is the interleaving of tree search and deep learning, which has proven very successful in board games like Chess, Shogi and Go.
2 code implementations • 23 May 2018 • Thomas M. Moerland, Joost Broekens, Aske Plaat, Catholijn M. Jonker
Asymmetric termination of search trees introduces a type of uncertainty for which the standard upper confidence bound (UCB) formula does not account.
no code implementations • 29 Nov 2017 • Thomas M. Moerland, Joost Broekens, Catholijn M. Jonker
This paper studies directed exploration for reinforcement learning agents by tracking uncertainty about the value of each available action.
no code implementations • 15 May 2017 • Thomas M. Moerland, Joost Broekens, Catholijn M. Jonker
This article provides the first survey of computational models of emotion in reinforcement learning (RL) agents.
1 code implementation • 1 May 2017 • Thomas M. Moerland, Joost Broekens, Catholijn M. Jonker
In this paper we study how to learn stochastic, multimodal transition dynamics in reinforcement learning (RL) tasks.
Model-based Reinforcement Learning reinforcement-learning +2
no code implementations • 8 Apr 2014 • Joost Broekens, Tim Baarslag
We found that overly optimistic risk perception (outcome optimism) results in risk taking and in persistent gambling behaviour in addition to high intensity of sensations.