Search Results for author: Antoine Grosnit

Found 15 papers, 7 papers with code

Many of Your DPOs are Secretly One: Attempting Unification Through Mutual Information

no code implementations2 Jan 2025 Rasul Tutnov, Antoine Grosnit, Haitham Bou-Ammar

However, the vast number of DPO variants in the literature has made it increasingly difficult for researchers to navigate and fully grasp the connections between these approaches.

Navigate

ShortCircuit: AlphaZero-Driven Circuit Design

no code implementations19 Aug 2024 Dimitrios Tsaras, Antoine Grosnit, Lei Chen, Zhiyao Xie, Haitham Bou-Ammar, Mingxuan Yuan

Chip design relies heavily on generating Boolean circuits, such as AND-Inverter Graphs (AIGs), from functional descriptions like truth tables.

ROS-LLM: A ROS framework for embodied AI with task feedback and structured reasoning

1 code implementation28 Jun 2024 Christopher E. Mower, Yuhui Wan, Hongzhan Yu, Antoine Grosnit, Jonas Gonzalez-Billandon, Matthieu Zimmer, Jinlong Wang, Xinyu Zhang, Yao Zhao, Anbang Zhai, Puze Liu, Daniel Palenicek, Davide Tateo, Cesar Cadena, Marco Hutter, Jan Peters, Guangjian Tian, Yuzheng Zhuang, Kun Shao, Xingyue Quan, Jianye Hao, Jun Wang, Haitham Bou-Ammar

Key features of the framework include: integration of ROS with an AI agent connected to a plethora of open-source and commercial LLMs, automatic extraction of a behavior from the LLM output and execution of ROS actions/services, support for three behavior modes (sequence, behavior tree, state machine), imitation learning for adding new robot actions to the library of possible actions, and LLM reflection via human and environment feedback.

AI Agent Imitation Learning

Why Can Large Language Models Generate Correct Chain-of-Thoughts?

no code implementations20 Oct 2023 Rasul Tutunov, Antoine Grosnit, Juliusz Ziomek, Jun Wang, Haitham Bou-Ammar

This paper delves into the capabilities of large language models (LLMs), specifically focusing on advancing the theoretical comprehension of chain-of-thought prompting.

Text Generation

Framework and Benchmarks for Combinatorial and Mixed-variable Bayesian Optimization

1 code implementation NeurIPS 2023 Kamil Dreczkowski, Antoine Grosnit, Haitham Bou Ammar

This paper introduces a modular framework for Mixed-variable and Combinatorial Bayesian Optimization (MCBO) to address the lack of systematic benchmarking and standardized evaluation in the field.

Bayesian Optimization Benchmarking

Contextual Causal Bayesian Optimisation

no code implementations29 Jan 2023 Vahan Arsenyan, Antoine Grosnit, Haitham Bou-Ammar

Causal Bayesian optimisation (CaBO) combines causality with Bayesian optimisation (BO) and shows that there are situations where the optimal reward is not achievable if causal knowledge is ignored.

Bayesian Optimisation Multi-Armed Bandits

Sample-Efficient Optimisation with Probabilistic Transformer Surrogates

no code implementations27 May 2022 Alexandre Maraval, Matthieu Zimmer, Antoine Grosnit, Rasul Tutunov, Jun Wang, Haitham Bou Ammar

First, we notice that these models are trained on uniformly distributed inputs, which impairs predictive accuracy on non-uniform data - a setting arising from any typical BO loop due to exploration-exploitation trade-offs.

Bayesian Optimisation Gaussian Processes

BOiLS: Bayesian Optimisation for Logic Synthesis

no code implementations11 Nov 2021 Antoine Grosnit, Cedric Malherbe, Rasul Tutunov, Xingchen Wan, Jun Wang, Haitham Bou Ammar

Optimising the quality-of-results (QoR) of circuits during logic synthesis is a formidable challenge necessitating the exploration of exponentially sized search spaces.

Bayesian Optimisation Navigate

Decentralized Deterministic Multi-Agent Reinforcement Learning

no code implementations19 Feb 2021 Antoine Grosnit, Desmond Cai, Laura Wynter

We extend those results to offer a provably-convergent decentralized actor-critic algorithm for learning deterministic policies on continuous action spaces.

Multi-agent Reinforcement Learning reinforcement-learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.