no code implementations • 10 Feb 2025 • Jordi Armengol-Estapé, Quentin Carbonneaux, Tianjun Zhang, Aram H. Markosyan, Volker Seeker, Chris Cummins, Melanie Kambadur, Michael F. P. O'Boyle, Sida Wang, Gabriel Synnaeve, Hugh James Leather
Code generation and understanding are critical capabilities for large language models (LLMs).
no code implementations • 31 Dec 2024 • Davide Italiano, Chris Cummins
To date we have reported 24 confirmed bugs in production compilers, and conclude that LLM-assisted testing is a promising avenue for detecting optimization bugs in real world compilers.
no code implementations • 11 Oct 2024 • Chris Cummins, Volker Seeker, Jordi Armengol-Estapé, Aram H. Markosyan, Gabriel Synnaeve, Hugh Leather
Unlike the direct rewrite approach, LLM-generated transformations are easy to inspect, debug, and validate.
no code implementations • 27 Jun 2024 • Chris Cummins, Volker Seeker, Dejan Grubisic, Baptiste Roziere, Jonas Gehring, Gabriel Synnaeve, Hugh Leather
To address this gap, we introduce Meta Large Language Model Compiler (LLM Compiler), a suite of robust, openly available, pre-trained models specifically designed for code optimization tasks.
no code implementations • 18 Mar 2024 • Dejan Grubisic, Chris Cummins, Volker Seeker, Hugh Leather
We introduce a novel paradigm in compiler optimization powered by Large Language Models with compiler feedback to optimize the code size of LLVM assembly.
no code implementations • 28 Feb 2024 • Dejan Grubisic, Chris Cummins, Volker Seeker, Hugh Leather
Large language models show great potential in generating and optimizing code.
no code implementations • 11 Sep 2023 • Chris Cummins, Volker Seeker, Dejan Grubisic, Mostafa Elhoushi, Youwei Liang, Baptiste Roziere, Jonas Gehring, Fabian Gloeckle, Kim Hazelwood, Gabriel Synnaeve, Hugh Leather
We explore the novel application of Large Language Models to code optimization.
no code implementations • 4 Sep 2023 • Dejan Grubisic, Bram Wasti, Chris Cummins, John Mellor-Crummey, Aleksandar Zlateski
Advanced compiler technology is crucial for enabling machine learning applications to run on novel hardware, but traditional compilers fail to deliver performance, popular auto-tuners have long search times and expert-optimized libraries introduce unsustainable costs.
no code implementations • 21 May 2023 • Jordi Armengol-Estapé, Jackson Woodruff, Chris Cummins, Michael F. P. O'Boyle
SLaDe is up to 6 times more accurate than Ghidra, a state-of-the-art, industrial-strength decompiler and up to 4 times more accurate than the large language model ChatGPT and generates significantly more readable code than both.
no code implementations • 2 Mar 2023 • Foivos Tsimpourlas, Pavlos Petoumenos, Min Xu, Chris Cummins, Kim Hazelwood, Ajitha Rajan, Hugh Leather
We improve this with BenchDirect which utilizes a directed LM that infills programs by jointly observing source code context and the compiler features that are targeted.
no code implementations • 9 Jan 2023 • Youwei Liang, Kevin Stone, Ali Shameli, Chris Cummins, Mostafa Elhoushi, Jiadong Guo, Benoit Steiner, Xiaomeng Yang, Pengtao Xie, Hugh Leather, Yuandong Tian
Finding the optimal pass sequence of compilation can lead to a significant reduction in program size and/or improvement in program efficiency.
1 code implementation • 13 Aug 2022 • Foivos Tsimpourlas, Pavlos Petoumenos, Min Xu, Chris Cummins, Kim Hazelwood, Ajitha Rajan, Hugh Leather
We develop BenchPress, the first ML benchmark generator for compilers that is steerable within feature space representations of source code.
1 code implementation • 24 Dec 2021 • Nadav Rotem, Chris Cummins
We perform offline training using information that is collected from a large corpus of binaries that have branch probabilities information.
1 code implementation • 17 Sep 2021 • Chris Cummins, Bram Wasti, Jiadong Guo, Brandon Cui, Jason Ansel, Sahir Gomez, Somya Jain, Jia Liu, Olivier Teytaud, Benoit Steiner, Yuandong Tian, Hugh Leather
What is needed is an easy, reusable experimental infrastructure for real world compiler optimization tasks that can serve as a common benchmark for comparing techniques, and as a platform to accelerate progress in the field.
2 code implementations • NeurIPS 2021 • Kevin Yang, Tianjun Zhang, Chris Cummins, Brandon Cui, Benoit Steiner, Linnan Wang, Joseph E. Gonzalez, Dan Klein, Yuandong Tian
Path planning, the problem of efficiently discovering high-reward trajectories, often requires optimizing a high-dimensional and multimodal reward function.
no code implementations • 30 Nov 2020 • Benoit Steiner, Chris Cummins, Horace He, Hugh Leather
As machine learning techniques become ubiquitous, the efficiency of neural network implementations is becoming correspondingly paramount.
no code implementations • 21 Nov 2020 • Chris Cummins, Hugh Leather, Zacharias Fisches, Tal Ben-Nun, Torsten Hoefler, Michael O'Boyle
Compiler architects increasingly look to machine learning when building heuristics for compiler optimization.
2 code implementations • 23 Mar 2020 • Chris Cummins, Zacharias V. Fisches, Tal Ben-Nun, Torsten Hoefler, Hugh Leather
We introduce ProGraML - Program Graphs for Machine Learning - a novel graph-based program representation using a low level, language agnostic, and portable format; and machine learning models capable of performing complex downstream tasks over these graphs.
1 code implementation • 2017 26th International Conference on Parallel Architectures and Compilation Techniques (PACT) 2017 • Chris Cummins, Pavlos Petoumenos, Zheng Wang, Hugh Leather
We develop a deep neural network that learns heuristics over raw code, entirely without using code features.
1 code implementation • 8 Nov 2015 • Chris Cummins, Pavlos Petoumenos, Michel Steuwer, Hugh Leather
Selecting an appropriate workgroup size is critical for the performance of OpenCL kernels, and requires knowledge of the underlying hardware, the data being operated on, and the implementation of the kernel.
Distributed, Parallel, and Cluster Computing