Search Results for author: Pierre-Yves Oudeyer

Found 77 papers, 27 papers with code

Inferring the Phylogeny of Large Language Models and Predicting their Performances in Benchmarks

no code implementations6 Apr 2024 Nicolas Yax, Pierre-Yves Oudeyer, Stefano Palminteri

This paper introduces PhyloLM, a method applying phylogenetic algorithms to Large Language Models to explore their finetuning relationships, and predict their performance characteristics.

Cultural evolution in populations of Large Language Models

1 code implementation13 Mar 2024 Jérémy Perez, Corentin Léger, Marcela Ovando-Tellez, Chris Foulon, Joan Dussauld, Pierre-Yves Oudeyer, Clément Moulin-Frier

We here present a framework for simulating cultural evolution in populations of LLMs, allowing the manipulation of variables known to be important in cultural evolution, such as network structure, personality, and the way social information is aggregated and transformed.

Stick to your Role! Stability of Personal Values Expressed in Large Language Models

no code implementations19 Feb 2024 Grgur Kovač, Rémy Portelas, Masataka Sawayama, Peter Ford Dominey, Pierre-Yves Oudeyer

In this paper, we present a case-study about the stability of value expression over different contexts (simulated conversations on different topics), and as measured using a standard psychology questionnaire (PVQ) and a behavioral downstream task.

Multiple-choice

Discovering Sensorimotor Agency in Cellular Automata using Diversity Search

no code implementations14 Feb 2024 Gautier Hamon, Mayalen Etcheverry, Bert Wang-Chak Chan, Clément Moulin-Frier, Pierre-Yves Oudeyer

The research field of Artificial Life studies how life-like phenomena such as autopoiesis, agency, or self-regulation can self-organize in computer simulations.

Artificial Life Navigate

Improved Performances and Motivation in Intelligent Tutoring Systems: Combining Machine Learning and Learner Choice

no code implementations16 Jan 2024 Benjamin Clément, Hélène Sauzéon, Didier Roy, Pierre-Yves Oudeyer

In this context, the ZPDES algorithm, based on the Learning Progress Hypothesis (LPH) and multi-armed bandit machine learning techniques, sequences exercises that maximize learning progress (LP).

Meta-Diversity Search in Complex Systems, A Recipe for Artificial Open-Endedness ?

no code implementations1 Dec 2023 Mayalen Etcheverry, Bert Wang-Chak Chan, Clément Moulin-Frier, Pierre-Yves Oudeyer

Holmes incrementally learns a hierarchy of modular representations to characterize divergent sources of diversity and uses a goal-based intrinsically-motivated exploration as the diversity search strategy.

SBMLtoODEjax: Efficient Simulation and Optimization of Biological Network Models in JAX

1 code implementation17 Jul 2023 Mayalen Etcheverry, Michael Levin, Clément Moulin-Frier, Pierre-Yves Oudeyer

Advances in bioengineering and biomedicine demand a deep understanding of the dynamic behavior of biological systems, ranging from protein pathways to complex cellular processes.

Large Language Models as Superpositions of Cultural Perspectives

no code implementations15 Jul 2023 Grgur Kovač, Masataka Sawayama, Rémy Portelas, Cédric Colas, Peter Ford Dominey, Pierre-Yves Oudeyer

We introduce the concept of perspective controllability, which refers to a model's affordance to adopt various perspectives with differing values and personality traits.

The SocialAI School: Insights from Developmental Psychology Towards Artificial Socio-Cultural Agents

no code implementations15 Jul 2023 Grgur Kovač, Rémy Portelas, Peter Ford Dominey, Pierre-Yves Oudeyer

Developmental psychologists have long-established the importance of socio-cognitive abilities in human intelligence.

Augmenting Autotelic Agents with Large Language Models

no code implementations21 May 2023 Cédric Colas, Laetitia Teodorescu, Pierre-Yves Oudeyer, Xingdi Yuan, Marc-Alexandre Côté

Without relying on any hand-coded goal representations, reward functions or curriculum, we show that LMA3 agents learn to master a large diversity of skills in a task-agnostic text-based environment.

Common Sense Reasoning Language Modelling

Supporting Qualitative Analysis with Large Language Models: Combining Codebook with GPT-3 for Deductive Coding

no code implementations17 Apr 2023 Ziang Xiao, Xingdi Yuan, Q. Vera Liao, Rania Abdelghani, Pierre-Yves Oudeyer

In this study, we explored the use of large language models (LLMs) in supporting deductive coding, a major category of qualitative analysis where researchers use pre-determined codebooks to label the data into a fixed set of codes.

A Song of Ice and Fire: Analyzing Textual Autotelic Agents in ScienceWorld

no code implementations10 Feb 2023 Laetitia Teodorescu, Xingdi Yuan, Marc-Alexandre Côté, Pierre-Yves Oudeyer

We show the importance of selectivity from the social peer's feedback; that experience replay needs to over-sample examples of rare goals; and that following self-generated goal sequences where the agent's competence is intermediate leads to significant improvements in final performance.

Grounding Large Language Models in Interactive Environments with Online Reinforcement Learning

3 code implementations6 Feb 2023 Thomas Carta, Clément Romac, Thomas Wolf, Sylvain Lamprier, Olivier Sigaud, Pierre-Yves Oudeyer

Using an interactive textual environment designed to study higher-level forms of functional grounding, and a set of spatial and navigation tasks, we study several scientific questions: 1) Can LLMs boost sample efficiency for online learning of various RL tasks?

Decision Making reinforcement-learning +1

Flow-Lenia: Towards open-ended evolution in cellular automata through mass conservation and parameter localization

1 code implementation14 Dec 2022 Erwan Plantec, Gautier Hamon, Mayalen Etcheverry, Pierre-Yves Oudeyer, Clément Moulin-Frier, Bert Wang-Chak Chan

Finally, we show that Flow Lenia enables the integration of the parameters of the CA update rules within the CA dynamics, making them dynamic and localized, allowing for multi-species simulations, with locally coherent update rules that define properties of the emerging creatures, and that can be mixed with neighbouring rules.

Artificial Life

GPT-3-driven pedagogical agents for training children's curious question-asking skills

no code implementations25 Nov 2022 Rania Abdelghani, Yen-Hsiang Wang, Xingdi Yuan, Tong Wang, Pauline Lucas, Hélène Sauzéon, Pierre-Yves Oudeyer

In this context, we propose to leverage advances in the natural language processing field (NLP) and investigate the efficiency of using a large language model (LLM) for automating the production of the pedagogical content of a curious question-asking (QA) training.

Language Modelling Large Language Model

Contrastive Multimodal Learning for Emergence of Graphical Sensory-Motor Communication

no code implementations3 Oct 2022 Tristan Karch, Yoann Lemesle, Romain Laroche, Clément Moulin-Frier, Pierre-Yves Oudeyer

In this paper, we investigate whether artificial agents can develop a shared language in an ecological setting where communication relies on a sensory-motor channel.

Automatic Exploration of Textual Environments with Language-Conditioned Autotelic Agents

no code implementations NAACL (Wordplay) 2022 Laetitia Teodorescu, Eric Yuan, Marc-Alexandre Côté, Pierre-Yves Oudeyer

In this extended abstract we discuss the opportunities and challenges of studying intrinsically-motivated agents for exploration in textual environments.

EAGER: Asking and Answering Questions for Automatic Reward Shaping in Language-guided RL

1 code implementation20 Jun 2022 Thomas Carta, Pierre-Yves Oudeyer, Olivier Sigaud, Sylvain Lamprier

Reinforcement learning (RL) in long horizon and sparse reward tasks is notoriously difficult and requires a lot of training steps.

Question Answering Question Generation +2

Social Network Structure Shapes Innovation: Experience-sharing in RL with SAPIENS

no code implementations10 Jun 2022 Eleni Nisioti, Mateo Mahaut, Pierre-Yves Oudeyer, Ida Momennejad, Clément Moulin-Frier

Comparing the level of innovation achieved by different social network structures across different tasks shows that, first, consistent with human findings, experience sharing within a dynamic structure achieves the highest level of innovation in tasks with a deceptive nature and large search spaces.

Cultural Vocal Bursts Intensity Prediction Reinforcement Learning (RL)

Language and Culture Internalisation for Human-Like Autotelic AI

no code implementations2 Jun 2022 Cédric Colas, Tristan Karch, Clément Moulin-Frier, Pierre-Yves Oudeyer

Building autonomous agents able to grow open-ended repertoires of skills across their lives is a fundamental goal of artificial intelligence (AI).

Attribute Cultural Vocal Bursts Intensity Prediction

Asking for Knowledge: Training RL Agents to Query External Knowledge Using Language

no code implementations12 May 2022 Iou-Jen Liu, Xingdi Yuan, Marc-Alexandre Côté, Pierre-Yves Oudeyer, Alexander G. Schwing

In order to study how agents can be taught to query external knowledge via language, we first introduce two new environments: the grid-world-based Q-BabyAI and the text-based Q-TextWorld.

Language-biased image classification: evaluation based on semantic representations

1 code implementation26 Jan 2022 Yoann Lemesle, Masataka Sawayama, Guillermo Valle-Perez, Maxime Adolphe, Hélène Sauzéon, Pierre-Yves Oudeyer

This suggests that the semantic word representation in the CLIP visual processing is not shared with the image representation, although the word representation strongly dominates for word-embedded images.

Classification Image Classification

Learning to Guide and to Be Guided in the Architect-Builder Problem

1 code implementation ICLR 2022 Paul Barde, Tristan Karch, Derek Nowrouzezahrai, Clément Moulin-Frier, Christopher Pal, Pierre-Yves Oudeyer

ABIG results in a low-level, high-frequency, guiding communication protocol that not only enables an architect-builder pair to solve the task at hand, but that can also generalize to unseen tasks.

Imitation Learning

Evaluating Language-biased image classification based on semantic compositionality

no code implementations ICLR 2022 Yoann Lemesle, Masataka Sawayama, Guillermo Valle-Perez, Maxime Adolphe, Hélène Sauzéon, Pierre-Yves Oudeyer

This suggests that the semantic word representation in the CLIP visual processing is not shared with the image representation, although the word representation strongly dominates for word-embedded images.

Classification Image Classification

SocialAI: Benchmarking Socio-Cognitive Abilities in Deep Reinforcement Learning Agents

no code implementations2 Jul 2021 Grgur Kovač, Rémy Portelas, Katja Hofmann, Pierre-Yves Oudeyer

In this paper, we argue that aiming towards human-level AI requires a broader set of key social skills: 1) language use in complex and variable social contexts; 2) beyond language, complex embodied communication in multimodal settings within constantly evolving social worlds.

Benchmarking reinforcement-learning +1

Causal Reinforcement Learning using Observational and Interventional Data

1 code implementation28 Jun 2021 Maxime Gasse, Damien Grasset, Guillaume Gaudron, Pierre-Yves Oudeyer

We then ask the following questions: can the online and offline experiences be safely combined for learning a causal model ?

Causal Inference Model-based Reinforcement Learning +2

Transflower: probabilistic autoregressive dance generation with multimodal attention

no code implementations25 Jun 2021 Guillermo Valle-Pérez, Gustav Eje Henter, Jonas Beskow, André Holzapfel, Pierre-Yves Oudeyer, Simon Alexanderson

First, we present a novel probabilistic autoregressive architecture that models the distribution over future poses with a normalizing flow conditioned on previous poses as well as music context, using a multimodal transformer encoder.

Grounding Spatio-Temporal Language with Transformers

1 code implementation NeurIPS 2021 Tristan Karch, Laetitia Teodorescu, Katja Hofmann, Clément Moulin-Frier, Pierre-Yves Oudeyer

While there is an extended literature studying how machines can learn grounded language, the topic of how to learn spatio-temporal linguistic concepts is still largely uncharted.

Towards Teachable Autotelic Agents

no code implementations25 May 2021 Olivier Sigaud, Ahmed Akakzia, Hugo Caselles-Dupré, Cédric Colas, Pierre-Yves Oudeyer, Mohamed Chetouani

In the field of Artificial Intelligence, these extremes respectively map to autonomous agents learning from their own signals and interactive learning agents fully taught by their teachers.

SocialAI 0.1: Towards a Benchmark to Stimulate Research on Socio-Cognitive Abilities in Deep Reinforcement Learning Agents

no code implementations27 Apr 2021 Grgur Kovač, Rémy Portelas, Katja Hofmann, Pierre-Yves Oudeyer

Building embodied autonomous agents capable of participating in social interactions with humans is one of the main challenges in AI.

TeachMyAgent: a Benchmark for Automatic Curriculum Learning in Deep RL

1 code implementation17 Mar 2021 Clément Romac, Rémy Portelas, Katja Hofmann, Pierre-Yves Oudeyer

Training autonomous agents able to generalize to multiple tasks is a key target of Deep Reinforcement Learning (DRL) research.

Intelligent behavior depends on the ecological niche: Scaling up AI to human-like intelligence in socio-cultural environments

no code implementations11 Mar 2021 Manfred Eppe, Pierre-Yves Oudeyer

This paper outlines a perspective on the future of AI, discussing directions for machines models of human-like intelligence.

Autotelic Agents with Intrinsically Motivated Goal-Conditioned Reinforcement Learning: a Short Survey

no code implementations17 Dec 2020 Cédric Colas, Tristan Karch, Olivier Sigaud, Pierre-Yves Oudeyer

Developmental RL is concerned with the use of deep RL algorithms to tackle a developmental problem -- the $intrinsically$ $motivated$ $acquisition$ $of$ $open$-$ended$ $repertoires$ $of$ $skills$.

reinforcement-learning Reinforcement Learning (RL)

Meta Automatic Curriculum Learning

no code implementations16 Nov 2020 Rémy Portelas, Clément Romac, Katja Hofmann, Pierre-Yves Oudeyer

In such complex task spaces, it is essential to rely on some form of Automatic Curriculum Learning (ACL) to adapt the task sampling distribution to a given learning agent, instead of randomly sampling tasks, as many could end up being either trivial or unfeasible.

EpidemiOptim: A Toolbox for the Optimization of Control Policies in Epidemiological Models

2 code implementations9 Oct 2020 Cédric Colas, Boris Hejblum, Sébastien Rouillon, Rodolphe Thiébaut, Pierre-Yves Oudeyer, Clément Moulin-Frier, Mélanie Prague

Epidemiologists model the dynamics of epidemics in order to propose control strategies based on pharmaceutical and non-pharmaceutical interventions (contact limitation, lock down, vaccination, etc).

Epidemiology Evolutionary Algorithms +5

GRIMGEP: Learning Progress for Robust Goal Sampling in Visual Deep Reinforcement Learning

no code implementations10 Aug 2020 Grgur Kovač, Adrien Laversanne-Finot, Pierre-Yves Oudeyer

However, a currently known limitation of agents trying to maximize the diversity of sampled goals is that they tend to get attracted to noise or more generally to parts of the environments that cannot be controlled (distractors).

reinforcement-learning Reinforcement Learning (RL)

Language-Goal Imagination to Foster Creative Exploration in Deep RL

no code implementations ICML Workshop LaReL 2020 Tristan Karch, Nicolas Lair, Cédric Colas, Jean-Michel Dussoux, Clément Moulin-Frier, Peter Ford Dominey, Pierre-Yves Oudeyer

We introduce the Playground environment and study how this form of goal imagination improves generalization and exploration over agents lacking this capacity.

Grounding Language to Autonomously-Acquired Skills via Goal Generation

1 code implementation ICLR 2021 Ahmed Akakzia, Cédric Colas, Pierre-Yves Oudeyer, Mohamed Chetouani, Olivier Sigaud

In a second stage (L -> G), it trains a language-conditioned goal generator to generate semantic goals that match the constraints expressed in language-based inputs.

Language Acquisition

Progressive growing of self-organized hierarchical representations for exploration

no code implementations13 May 2020 Mayalen Etcheverry, Pierre-Yves Oudeyer, Chris Reinke

A central challenge is how to learn incrementally representations in order to progressively build a map of the discovered structures and re-use it to further explore.

Representation Learning

SpatialSim: Recognizing Spatial Configurations of Objects with Graph Neural Networks

no code implementations9 Apr 2020 Laetitia Teodorescu, Katja Hofmann, Pierre-Yves Oudeyer

Recognizing precise geometrical configurations of groups of objects is a key capability of human spatial cognition, yet little studied in the deep learning literature so far.

Trying AGAIN instead of Trying Longer: Prior Learning for Automatic Curriculum Learning

no code implementations7 Apr 2020 Rémy Portelas, Katja Hofmann, Pierre-Yves Oudeyer

A major challenge in the Deep RL (DRL) community is to train agents able to generalize over unseen situations, which is often approached by training them on a diversity of tasks (or environments).

Deep Sets for Generalization in RL

no code implementations20 Mar 2020 Tristan Karch, Cédric Colas, Laetitia Teodorescu, Clément Moulin-Frier, Pierre-Yves Oudeyer

This paper investigates the idea of encoding object-centered representations in the design of the reward function and policy architectures of a language-guided reinforcement learning agent.

Navigate Object +3

Automatic Curriculum Learning For Deep RL: A Short Survey

no code implementations10 Mar 2020 Rémy Portelas, Cédric Colas, Lilian Weng, Katja Hofmann, Pierre-Yves Oudeyer

Automatic Curriculum Learning (ACL) has become a cornerstone of recent successes in Deep Reinforcement Learning (DRL). These methods shape the learning trajectories of agents by challenging them with tasks adapted to their capacities.

reinforcement-learning Reinforcement Learning (RL)

Multi-Agent Reinforcement Learning as a Computational Tool for Language Evolution Research: Historical Context and Future Challenges

no code implementations20 Feb 2020 Clément Moulin-Frier, Pierre-Yves Oudeyer

Computational models of emergent communication in agent populations are currently gaining interest in the machine learning community due to recent advances in Multi-Agent Reinforcement Learning (MARL).

BIG-bench Machine Learning Multi-agent Reinforcement Learning +3

User-in-the-loop Adaptive Intent Detection for Instructable Digital Assistant

1 code implementation16 Jan 2020 Nicolas Lair, Clément Delgrange, David Mugisha, Jean-Michel Dussoux, Pierre-Yves Oudeyer, Peter Ford Dominey

To provide such functionalities, NL interpretation in traditional assistants should be improved: (1) The intent identification system should be able to recognize new forms of known intents, and to acquire new intents as they are expressed by the user.

Intent Detection Natural Language Understanding +3

Language Grounding through Social Interactions and Curiosity-Driven Multi-Goal Learning

no code implementations8 Nov 2019 Nicolas Lair, Cédric Colas, Rémy Portelas, Jean-Michel Dussoux, Peter Ford Dominey, Pierre-Yves Oudeyer

We propose LE2 (Language Enhanced Exploration), a learning algorithm leveraging intrinsic motivations and natural language (NL) interactions with a descriptive social partner (SP).

Active Learning Descriptive

Teacher algorithms for curriculum learning of Deep RL in continuously parameterized environments

2 code implementations16 Oct 2019 Rémy Portelas, Cédric Colas, Katja Hofmann, Pierre-Yves Oudeyer

We consider the problem of how a teacher algorithm can enable an unknown Deep Reinforcement Learning (DRL) student to become good at a skill over a wide range of diverse environments.

Intrinsically Motivated Discovery of Diverse Patterns in Self-Organizing Systems

no code implementations ICLR 2020 Chris Reinke, Mayalen Etcheverry, Pierre-Yves Oudeyer

Using a continuous GOL as a testbed, we show that recent intrinsically-motivated machine learning algorithms (POP-IMGEPs), initially developed for learning of inverse models in robotics, can be transposed and used in this novel application area.

Autonomous Goal Exploration using Learned Goal Spaces for Visuomotor Skill Acquisition in Robots

no code implementations10 Jun 2019 Adrien Laversanne-Finot, Alexandre Péré, Pierre-Yves Oudeyer

The automatic and efficient discovery of skills, without supervision, for long-living autonomous agents, remains a challenge of Artificial Intelligence.

Representation Learning

A Hitchhiker's Guide to Statistical Comparisons of Reinforcement Learning Algorithms

2 code implementations15 Apr 2019 Cédric Colas, Olivier Sigaud, Pierre-Yves Oudeyer

Consistently checking the statistical significance of experimental results is the first mandatory step towards reproducible science.

reinforcement-learning Reinforcement Learning (RL)

Computational and Robotic Models of Early Language Development: A Review

no code implementations25 Mar 2019 Pierre-Yves Oudeyer, George Kachergis, William Schueller

Keywords: Early language learning, Computational and robotic models, machine learning, development, embodiment, social interaction, intrinsic motivation, self-organization, dynamical systems, complexity.

BIG-bench Machine Learning Language Acquisition

CURIOUS: Intrinsically Motivated Modular Multi-Goal Reinforcement Learning

1 code implementation15 Oct 2018 Cédric Colas, Pierre Fournier, Olivier Sigaud, Mohamed Chetouani, Pierre-Yves Oudeyer

In open-ended environments, autonomous learning agents must set their own goals and build their own curriculum through an intrinsically motivated exploration.

Efficient Exploration Multi-Goal Reinforcement Learning +2

Diversity-Driven Selection of Exploration Strategies in Multi-Armed Bandits

no code implementations23 Aug 2018 Fabien C. Y. Benureau, Pierre-Yves Oudeyer

We consider a scenario where an agent has multiple available strategies to explore an unknown environment.

Multi-Armed Bandits

Curiosity Driven Exploration of Learned Disentangled Goal Spaces

1 code implementation4 Jul 2018 Adrien Laversanne-Finot, Alexandre Péré, Pierre-Yves Oudeyer

Finally, we show that the measure of learning progress, used to drive curiosity-driven exploration, can be used simultaneously to discover abstract independently controllable features of the environment.

Efficient Exploration Representation Learning

Accuracy-based Curriculum Learning in Deep Reinforcement Learning

2 code implementations25 Jun 2018 Pierre Fournier, Olivier Sigaud, Mohamed Chetouani, Pierre-Yves Oudeyer

In this paper, we investigate a new form of automated curriculum learning based on adaptive selection of accuracy requirements, called accuracy-based curriculum learning.

reinforcement-learning Reinforcement Learning (RL)

How Many Random Seeds? Statistical Power Analysis in Deep Reinforcement Learning Experiments

1 code implementation21 Jun 2018 Cédric Colas, Olivier Sigaud, Pierre-Yves Oudeyer

Consistently checking the statistical significance of experimental results is one of the mandatory methodological steps to address the so-called "reproducibility crisis" in deep reinforcement learning.

reinforcement-learning Reinforcement Learning (RL)

Complexity Reduction in the Negotiation of New Lexical Conventions

1 code implementation15 May 2018 William Schueller, Vittorio Loreto, Pierre-Yves Oudeyer

We define here a new principled measure and a new strategy, based on the beliefs of each agent on the global state of the population.

Socially Guided Intrinsic Motivation for Robot Learning of Motor Skills

no code implementations19 Apr 2018 Sao Mai Nguyen, Pierre-Yves Oudeyer

This paper presents a technical approach to robot learning of motor skills which combines active intrinsically motivated learning with imitation learning.

Imitation Learning

Active choice of teachers, learning strategies and goals for a socially guided intrinsic motivation learner

no code implementations18 Apr 2018 Sao Mai Nguyen, Pierre-Yves Oudeyer

It has to choose actively and hierarchically at each learning episode: 1) what to learn: which outcome is most interesting to select as a goal to focus on for goal-directed exploration; 2) how to learn: which data collection strategy to use among self-exploration, mimicry and emulation; 3) once he has decided when and what to imitate by choosing mimicry or emulation, then he has to choose who to imitate, from a set of different teachers.

Active Learning

Unsupervised Learning of Goal Spaces for Intrinsically Motivated Goal Exploration

1 code implementation ICLR 2018 Alexandre Péré, Sébastien Forestier, Olivier Sigaud, Pierre-Yves Oudeyer

Intrinsically motivated goal exploration algorithms enable machines to discover repertoires of policies that produce a diversity of effects in complex environments.

Representation Learning

GEP-PG: Decoupling Exploration and Exploitation in Deep Reinforcement Learning Algorithms

1 code implementation ICML 2018 Cédric Colas, Olivier Sigaud, Pierre-Yves Oudeyer

In continuous action domains, standard deep reinforcement learning algorithms like DDPG suffer from inefficient exploration when facing sparse or deceptive reward problems.

reinforcement-learning Reinforcement Learning (RL)

Proximodistal Exploration in Motor Learning as an Emergent Property of Optimization

no code implementations14 Dec 2017 Freek Stulp, Pierre-Yves Oudeyer

Here, we formulate and study computationally the hypothesis that such patterns can emerge spontaneously as the result of a family of stochastic optimization processes (evolution strategies with covariance-matrix adaptation), without an innate encoding of a maturational schedule.

Stochastic Optimization

Autonomous development and learning in artificial intelligence and robotics: Scaling up deep learning to human--like learning

no code implementations5 Dec 2017 Pierre-Yves Oudeyer

Autonomous lifelong development and learning is a fundamental capability of humans, differentiating them from current deep learning systems.

Intrinsically Motivated Goal Exploration Processes with Automatic Curriculum Learning

3 code implementations7 Aug 2017 Sébastien Forestier, Rémy Portelas, Yoan Mollard, Pierre-Yves Oudeyer

We present an algorithmic approach called Intrinsically Motivated Goal Exploration Processes (IMGEP) to enable similar properties of autonomous learning in machines.

Developmental Learning Multi-Goal Reinforcement Learning +1

Open challenges in understanding development and evolution of speech forms: The roles of embodied self-organization, motivation and active exploration

no code implementations5 Jan 2016 Pierre-Yves Oudeyer

This article discusses open scientific challenges for understanding development and evolution of speech forms, as a commentary to Moulin-Frier et al. (Moulin-Frier et al., 2015).

What do we learn about development from baby robots?

no code implementations20 Jan 2015 Pierre-Yves Oudeyer

To address this challenge, building and experimenting with robots modeling the growing infant brain and body is crucial.

Multi-Armed Bandits for Intelligent Tutoring Systems

no code implementations11 Oct 2013 Benjamin Clement, Didier Roy, Pierre-Yves Oudeyer, Manuel Lopes

We present an approach to Intelligent Tutoring Systems which adaptively personalizes sequences of learning activities to maximize skills acquired by students, taking into account the limited time and motivational resources.

Multi-Armed Bandits

Active Learning of Inverse Models with Intrinsically Motivated Goal Exploration in Robots

1 code implementation21 Jan 2013 Adrien Baranes, Pierre-Yves Oudeyer

The architecture makes the robot sample actively novel parameterized tasks in the task space, based on a measure of competence progress, each of which triggers low-level goal-directed learning of the motor policy pa- rameters that allow to solve it.

Active Learning

Exploration in Model-based Reinforcement Learning by Empirically Estimating Learning Progress

no code implementations NeurIPS 2012 Manuel Lopes, Tobias Lang, Marc Toussaint, Pierre-Yves Oudeyer

Formal exploration approaches in model-based reinforcement learning estimate the accuracy of the currently learned model without consideration of the empirical prediction error.

Model-based Reinforcement Learning reinforcement-learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.