Search Results for author: Pablo Tano

Found 2 papers, 0 papers with code

A Local Temporal Difference Code for Distributional Reinforcement Learning

no code implementations NeurIPS 2020 Pablo Tano, Peter Dayan, Alexandre Pouget

Recent theoretical and experimental results suggest that the dopamine system implements distributional temporal difference backups, allowing learning of the entire distributions of the long-run values of states rather than just their expected values.

Distributional Reinforcement Learning Imputation +2

Towards a more flexible Language of Thought: Bayesian grammar updates after each concept exposure

no code implementations17 May 2018 Pablo Tano, Sergio Romano, Mariano Sigman, Alejo Salles, Santiago Figueira

Recent approaches to human concept learning have successfully combined the power of symbolic, infinitely productive rule systems and statistical learning to explain our ability to learn new concepts from just a few examples.

Descriptive

Cannot find the paper you are looking for? You can Submit a new open access paper.