Search Results for author: Haitham Bou-Ammar

Found 25 papers, 8 papers with code

Al-Khwarizmi: Discovering Physical Laws with Foundation Models

no code implementations3 Feb 2025 Christopher E. Mower, Haitham Bou-Ammar

Inferring physical laws from data is a central challenge in science and engineering, including but not limited to healthcare, physical sciences, biosciences, social sciences, sustainability, climate, and robotics.

RAG

Many of Your DPOs are Secretly One: Attempting Unification Through Mutual Information

no code implementations2 Jan 2025 Rasul Tutnov, Antoine Grosnit, Haitham Bou-Ammar

However, the vast number of DPO variants in the literature has made it increasingly difficult for researchers to navigate and fully grasp the connections between these approaches.

Navigate

ShortCircuit: AlphaZero-Driven Circuit Design

no code implementations19 Aug 2024 Dimitrios Tsaras, Antoine Grosnit, Lei Chen, Zhiyao Xie, Haitham Bou-Ammar, Mingxuan Yuan

Chip design relies heavily on generating Boolean circuits, such as AND-Inverter Graphs (AIGs), from functional descriptions like truth tables.

Human-like Episodic Memory for Infinite Context LLMs

1 code implementation12 Jul 2024 Zafeirios Fountas, Martin A Benfeghoul, Adnan Oomerjee, Fenia Christopoulou, Gerasimos Lampouras, Haitham Bou-Ammar, Jun Wang

Large language models (LLMs) have shown remarkable capabilities, but still struggle with processing extensive contexts, limiting their ability to maintain coherence and accuracy over long sequences.

Computational Efficiency Event Segmentation +2

ROS-LLM: A ROS framework for embodied AI with task feedback and structured reasoning

1 code implementation28 Jun 2024 Christopher E. Mower, Yuhui Wan, Hongzhan Yu, Antoine Grosnit, Jonas Gonzalez-Billandon, Matthieu Zimmer, Jinlong Wang, Xinyu Zhang, Yao Zhao, Anbang Zhai, Puze Liu, Daniel Palenicek, Davide Tateo, Cesar Cadena, Marco Hutter, Jan Peters, Guangjian Tian, Yuzheng Zhuang, Kun Shao, Xingyue Quan, Jianye Hao, Jun Wang, Haitham Bou-Ammar

Key features of the framework include: integration of ROS with an AI agent connected to a plethora of open-source and commercial LLMs, automatic extraction of a behavior from the LLM output and execution of ROS actions/services, support for three behavior modes (sequence, behavior tree, state machine), imitation learning for adding new robot actions to the library of possible actions, and LLM reflection via human and environment feedback.

AI Agent Imitation Learning

Safe Reinforcement Learning on the Constraint Manifold: Theory and Applications

no code implementations13 Apr 2024 Puze Liu, Haitham Bou-Ammar, Jan Peters, Davide Tateo

Indeed, safety specifications, often represented as constraints, can be complex and non-linear, making safety challenging to guarantee in learning systems.

reinforcement-learning Reinforcement Learning +1

Bayesian Reward Models for LLM Alignment

no code implementations20 Feb 2024 Adam X. Yang, Maxime Robeyns, Thomas Coste, Zhengyan Shi, Jun Wang, Haitham Bou-Ammar, Laurence Aitchison

To ensure that large language model (LLM) responses are helpful and non-toxic, a reward model trained on human preference data is usually used.

Language Modeling Language Modelling +1

Why Can Large Language Models Generate Correct Chain-of-Thoughts?

no code implementations20 Oct 2023 Rasul Tutunov, Antoine Grosnit, Juliusz Ziomek, Jun Wang, Haitham Bou-Ammar

This paper delves into the capabilities of large language models (LLMs), specifically focusing on advancing the theoretical comprehension of chain-of-thought prompting.

Text Generation

Are Random Decompositions all we need in High Dimensional Bayesian Optimisation?

2 code implementations30 Jan 2023 Juliusz Ziomek, Haitham Bou-Ammar

Learning decompositions of expensive-to-evaluate black-box functions promises to scale Bayesian optimisation (BO) to high-dimensional problems.

All Bayesian Optimisation

Contextual Causal Bayesian Optimisation

no code implementations29 Jan 2023 Vahan Arsenyan, Antoine Grosnit, Haitham Bou-Ammar

Causal Bayesian optimisation (CaBO) combines causality with Bayesian optimisation (BO) and shows that there are situations where the optimal reward is not achievable if causal knowledge is ignored.

Bayesian Optimisation Multi-Armed Bandits

Fast Kinodynamic Planning on the Constraint Manifold with Deep Neural Networks

1 code implementation11 Jan 2023 Piotr Kicki, Puze Liu, Davide Tateo, Haitham Bou-Ammar, Krzysztof Walas, Piotr Skrzypczyński, Jan Peters

Motion planning is a mature area of research in robotics with many well-established methods based on optimization or sampling the state space, suitable for solving kinematic motion planning.

Motion Planning

Reinforcement Learning in Presence of Discrete Markovian Context Evolution

no code implementations ICLR 2022 Hang Ren, Aivar Sootla, Taher Jafferjee, Junxiao Shen, Jun Wang, Haitham Bou-Ammar

We consider a context-dependent Reinforcement Learning (RL) setting, which is characterized by: a) an unknown finite number of not directly observable contexts; b) abrupt (discontinuous) context changes occurring during an episode; and c) Markovian context evolution.

reinforcement-learning Reinforcement Learning +2

Saute RL: Almost Surely Safe Reinforcement Learning Using State Augmentation

1 code implementation14 Feb 2022 Aivar Sootla, Alexander I. Cowen-Rivers, Taher Jafferjee, Ziyan Wang, David Mguni, Jun Wang, Haitham Bou-Ammar

Satisfying safety constraints almost surely (or with probability one) can be critical for the deployment of Reinforcement Learning (RL) in real-life applications.

reinforcement-learning Reinforcement Learning +2

Compositional ADAM: An Adaptive Compositional Solver

no code implementations10 Feb 2020 Rasul Tutunov, Minne Li, Alexander I. Cowen-Rivers, Jun Wang, Haitham Bou-Ammar

In this paper, we present C-ADAM, the first adaptive solver for compositional problems involving a non-linear functional nesting of expected values.

Meta-Learning

Learning High-level Representations from Demonstrations

no code implementations19 Feb 2018 Garrett Andersen, Peter Vrancx, Haitham Bou-Ammar

A common approach to HL, is to provide the agent with a number of high-level skills that solve small parts of the overall problem.

Montezuma's Revenge Open-Ended Question Answering +1

Balancing Two-Player Stochastic Games with Soft Q-Learning

no code implementations9 Feb 2018 Jordi Grau-Moya, Felix Leibfried, Haitham Bou-Ammar

Within the context of video games the notion of perfectly rational agents can be undesirable as it leads to uninteresting situations, where humans face tough adversarial decision makers.

Q-Learning Reinforcement Learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.