Search Results for author: Arthur Aubret

Found 14 papers, 4 papers with code

Human Gaze Boosts Object-Centered Representation Learning

no code implementations6 Jan 2025 Timothy Schaumlöffel, Arthur Aubret, Gemma Roig, Jochen Triesch

To account for the importance of central vision in humans, we crop the visual area around the gaze location.

Gaze Prediction Object +2

Seeing the Whole in the Parts in Self-Supervised Representation Learning

no code implementations6 Jan 2025 Arthur Aubret, Céline Teulière, Jochen Triesch

Here, we propose a new way to model spatial co-occurrences by aligning local representations (before pooling) with a global image representation.

Representation Learning Self-Supervised Learning

Active Gaze Behavior Boosts Self-Supervised Object Learning

no code implementations4 Nov 2024 Zhengyang Yu, Arthur Aubret, Marcel C. Raabe, Jane Yang, Chen Yu, Jochen Triesch

In this work, we explore whether a bio inspired visual learning model can harness toddlers' gaze behavior during a play session to develop view-invariant object recognition.

Object Object Recognition +1

Self-supervised visual learning from interactions with objects

1 code implementation9 Jul 2024 Arthur Aubret, Céline Teulière, Jochen Triesch

In our analysis, we find that the observed improvement is associated with a better viewpoint-wise alignment of different objects from the same category.

Object Representation Learning +1

Learning Object Semantic Similarity with Self-Supervision

no code implementations19 Apr 2024 Arthur Aubret, Timothy Schaumlöffel, Gemma Roig, Jochen Triesch

To achieve this, the model exploits two distinct strategies: the visuo-language alignment ensures that different objects of the same category are represented similarly, whereas the temporal alignment leverages that objects from the same context are frequently seen in succession to make their representations more similar.

Object Semantic Similarity +2

Self-Supervised Learning of Color Constancy

1 code implementation11 Apr 2024 Markus R. Ernst, Francisco M. López, Arthur Aubret, Roland W. Fleming, Jochen Triesch

Color constancy (CC) describes the ability of the visual system to perceive an object as having a relatively constant color despite changes in lighting conditions.

Color Constancy Self-Supervised Learning

MIMo: A Multi-Modal Infant Model for Studying Cognitive Development

1 code implementation7 Dec 2023 Dominik Mattern, Pierre Schumacher, Francisco M. López, Marcel C. Raabe, Markus R. Ernst, Arthur Aubret, Jochen Triesch

Human intelligence and human consciousness emerge gradually during the process of cognitive development.

An information-theoretic perspective on intrinsic motivation in reinforcement learning: a survey

no code implementations19 Sep 2022 Arthur Aubret, Laetitia Matignon, Salima Hassas

The reinforcement learning (RL) research area is very active, with an important number of new contributions; especially considering the emergent field of deep RL (DRL).

reinforcement-learning Reinforcement Learning (RL)

Time to augment self-supervised visual representation learning

no code implementations27 Jul 2022 Arthur Aubret, Markus Ernst, Céline Teulière, Jochen Triesch

Specifically, our analyses reveal that: 1) 3-D object manipulations drastically improve the learning of object categories; 2) viewing objects against changing backgrounds is important for learning to discard background-related information from the latent representation.

Contrastive Learning Object +2

Embodied vision for learning object representations

no code implementations12 May 2022 Arthur Aubret, Céline Teulière, Jochen Triesch

During each play session the agent views an object in multiple orientations before turning its body to view another object.

Contrastive Learning Object +1

DisTop: Discovering a Topological representation to learn diverse and rewarding skills

no code implementations6 Jun 2021 Arthur Aubret, Laetitia Matignon, Salima Hassas

The optimal way for a deep reinforcement learning (DRL) agent to explore is to learn a set of skills that achieves a uniform distribution of states.

Deep Reinforcement Learning Hierarchical Reinforcement Learning +4

ELSIM: End-to-end learning of reusable skills through intrinsic motivation

no code implementations ICML Workshop LifelongML 2020 Arthur Aubret, Laetitia Matignon, Salima Hassas

Then we show that our approach can scale on more difficult MuJoCo environments in which our agent is able to build a representation of skills which improve over a baseline both transfer learning and exploration when rewards are sparse.

Developmental Learning MuJoCo +1

Cannot find the paper you are looking for? You can Submit a new open access paper.