no code implementations • 10 Sep 2023 • Muhammad Burhan Hafez, Stefan Wermter
Our approach learns action and intention embeddings from self-organization of the observed movement and effect parts of unlabeled demonstrations and a higher-level behavior embedding from self-organization of the joint action-intention embeddings.
1 code implementation • 3 May 2023 • Muhammad Burhan Hafez, Tilman Immisch, Tom Weber, Stefan Wermter
Our approach organizes stored transitions into a concise environment-model-like network of state-nodes and transition-edges, merging similar samples to reduce the memory size and increase pair-wise distance among samples, which increases the relevancy of each sample.
1 code implementation • 14 Apr 2023 • Jonas Matthies, Muhammad Burhan Hafez, Mostafa Kotb, Stefan Wermter
On the other hand, a Q function is used to provide a good long-term estimate.
1 code implementation • 14 Mar 2023 • Xufeng Zhao, Mengdi Li, Cornelius Weber, Muhammad Burhan Hafez, Stefan Wermter
However, it remains challenging to ground LLMs in multimodal sensory input and continuous action output, while enabling a robot to interact with its environment and acquire novel information as its policies unfold.
no code implementations • 9 Jan 2023 • Ozan Özdemir, Matthias Kerzel, Cornelius Weber, Jae Hee Lee, Muhammad Burhan Hafez, Patrick Bruns, Stefan Wermter
Only occasionally, a learning infant would receive a matching verbal description of an action it is committing, which is similar to supervised learning.
1 code implementation • 4 Aug 2022 • Xufeng Zhao, Cornelius Weber, Muhammad Burhan Hafez, Stefan Wermter
Sound is one of the most informative and abundant modalities in the real world while being robust to sense without contacts by small and cheap sensors that can be placed on mobile devices.
no code implementations • 9 Jul 2021 • Muhammad Burhan Hafez, Stefan Wermter
Task inference is made by finding the nearest behavior embedding to a demonstrated behavior, which is used together with the environment state as input to a multi-task policy trained with reinforcement learning to optimize performance over tasks.
1 code implementation • 10 Feb 2021 • Julien Scholz, Cornelius Weber, Muhammad Burhan Hafez, Stefan Wermter
Using a model of the environment, reinforcement learning agents can plan their future moves and achieve superhuman performance in board games like Chess, Shogi, and Go, while remaining relatively sample-efficient.
1 code implementation • 19 Apr 2020 • Muhammad Burhan Hafez, Cornelius Weber, Matthias Kerzel, Stefan Wermter
In this paper, we present a novel dual-system motor learning approach where a meta-controller arbitrates online between model-based and model-free decisions based on an estimate of the local reliability of the learned model.
no code implementations • 10 Oct 2019 • Muhammad Burhan Hafez, Cornelius Weber, Matthias Kerzel, Stefan Wermter
The learned models are used to generate imagined experiences, augmenting the training set of real experiences.
no code implementations • 5 May 2019 • Muhammad Burhan Hafez, Cornelius Weber, Matthias Kerzel, Stefan Wermter
Recent success in deep reinforcement learning for continuous control has been dominated by model-free approaches which, unlike model-based approaches, do not suffer from representational limitations in making assumptions about the world dynamics and model errors inevitable in complex domains.
no code implementations • 26 Oct 2018 • Muhammad Burhan Hafez, Cornelius Weber, Matthias Kerzel, Stefan Wermter
In this paper, we present a new intrinsically motivated actor-critic algorithm for learning continuous motor skills directly from raw visual input.