Search Results for author: Jonas Gehring

Found 13 papers, 8 papers with code

What Makes Large Language Models Reason in (Multi-Turn) Code Generation?

no code implementations10 Oct 2024 Kunhao Zheng, Juliette Decugis, Jonas Gehring, Taco Cohen, Benjamin Negrevergne, Gabriel Synnaeve

Prompting techniques such as chain-of-thought have established themselves as a popular vehicle for improving the outputs of large language models (LLMs).

Code Generation

RLEF: Grounding Code LLMs in Execution Feedback with Reinforcement Learning

no code implementations2 Oct 2024 Jonas Gehring, Kunhao Zheng, Jade Copet, Vegard Mella, Taco Cohen, Gabriel Synnaeve

Large language models (LLMs) deployed as agents solve user-specified tasks over multiple steps while keeping the required manual engagement to a minimum.

reinforcement-learning Reinforcement Learning

Meta Large Language Model Compiler: Foundation Models of Compiler Optimization

no code implementations27 Jun 2024 Chris Cummins, Volker Seeker, Dejan Grubisic, Baptiste Roziere, Jonas Gehring, Gabriel Synnaeve, Hugh Leather

To address this gap, we introduce Meta Large Language Model Compiler (LLM Compiler), a suite of robust, openly available, pre-trained models specifically designed for code optimization tasks.

Compiler Optimization Language Modeling +2

The Larger the Better? Improved LLM Code-Generation via Budget Reallocation

1 code implementation31 Mar 2024 Michael Hassid, Tal Remez, Jonas Gehring, Roy Schwartz, Yossi Adi

On the other hand, in scenarios where unit-tests are unavailable, a ranking-based selection of candidates from the smaller model falls short of the performance of a single output from larger ones.

Code Generation

Code Llama: Open Foundation Models for Code

2 code implementations24 Aug 2023 Baptiste Rozière, Jonas Gehring, Fabian Gloeckle, Sten Sootla, Itai Gat, Xiaoqing Ellen Tan, Yossi Adi, Jingyu Liu, Romain Sauvestre, Tal Remez, Jérémy Rapin, Artyom Kozhevnikov, Ivan Evtimov, Joanna Bitton, Manish Bhatt, Cristian Canton Ferrer, Aaron Grattafiori, Wenhan Xiong, Alexandre Défossez, Jade Copet, Faisal Azhar, Hugo Touvron, Louis Martin, Nicolas Usunier, Thomas Scialom, Gabriel Synnaeve

We release Code Llama, a family of large language models for code based on Llama 2 providing state-of-the-art performance among open models, infilling capabilities, support for large input contexts, and zero-shot instruction following ability for programming tasks.

16k Code Generation +2

Leveraging Demonstrations with Latent Space Priors

1 code implementation26 Oct 2022 Jonas Gehring, Deepak Gopinath, Jungdam Won, Andreas Krause, Gabriel Synnaeve, Nicolas Usunier

Starting with a learned joint latent space, we separately train a generative model of demonstration sequences and an accompanying low-level policy.

Offline RL

Hierarchical Skills for Efficient Exploration

1 code implementation NeurIPS 2021 Jonas Gehring, Gabriel Synnaeve, Andreas Krause, Nicolas Usunier

We alleviate the need for prior knowledge by proposing a hierarchical skill learning framework that acquires skills of varying complexity in an unsupervised manner.

continuous-control Continuous Control +6

Forward Modeling for Partial Observation Strategy Games - A StarCraft Defogger

1 code implementation ICLR 2018 Gabriel Synnaeve, Zeming Lin, Jonas Gehring, Dan Gant, Vegard Mella, Vasil Khalidov, Nicolas Carion, Nicolas Usunier

We formulate the problem of defogging as state estimation and future state prediction from previous, partial observations in the context of real-time strategy games.

Decoder Starcraft

High-Level Strategy Selection under Partial Observability in StarCraft: Brood War

no code implementations21 Nov 2018 Jonas Gehring, Da Ju, Vegard Mella, Daniel Gant, Nicolas Usunier, Gabriel Synnaeve

We consider the problem of high-level strategy selection in the adversarial setting of real-time strategy games from a reinforcement learning perspective, where taking an action corresponds to switching to the respective strategy.

reinforcement-learning Reinforcement Learning +3

STARDATA: A StarCraft AI Research Dataset

1 code implementation7 Aug 2017 Zeming Lin, Jonas Gehring, Vasil Khalidov, Gabriel Synnaeve

We provide full game state data along with the original replays that can be viewed in StarCraft.

Diversity Imitation Learning +3

Cannot find the paper you are looking for? You can Submit a new open access paper.