Search Results for author: Daniel Tarlow

Found 44 papers, 16 papers with code

Experts Don't Cheat: Learning What You Don't Know By Predicting Pairs

no code implementations13 Feb 2024 Daniel D. Johnson, Daniel Tarlow, David Duvenaud, Chris J. Maddison

Identifying how much a model ${\widehat{p}}_{\theta}(Y|X)$ knows about the stochastic real-world process $p(Y|X)$ it was trained on is important to ensure it avoids producing incorrect or "hallucinated" answers or taking unsafe actions.

Image Classification Language Modelling +1

R-U-SURE? Uncertainty-Aware Code Suggestions By Maximizing Utility Across Random User Intents

1 code implementation1 Mar 2023 Daniel D. Johnson, Daniel Tarlow, Christian Walder

Large language models show impressive results at predicting structured text such as code, but also commonly introduce errors and hallucinations in their output.

A Library for Representing Python Programs as Graphs for Machine Learning

1 code implementation15 Aug 2022 David Bieber, Kensen Shi, Petros Maniatis, Charles Sutton, Vincent Hellendoorn, Daniel Johnson, Daniel Tarlow

Graph representations of programs are commonly a central element of machine learning for code research.

Learning to Improve Code Efficiency

no code implementations9 Aug 2022 Binghong Chen, Daniel Tarlow, Kevin Swersky, Martin Maas, Pablo Heiber, Ashish Naik, Milad Hashemi, Parthasarathy Ranganathan

To automatically learn these hints from the dataset, we propose a novel discrete variational auto-encoder, where each discrete latent variable represents a different learned category of code-edit that increases performance.

Repository-Level Prompt Generation for Large Language Models of Code

1 code implementation26 Jun 2022 Disha Shrivastava, Hugo Larochelle, Daniel Tarlow

With the success of large language models (LLMs) of code and their use as code assistants (e. g. Codex used in GitHub Copilot), techniques for introducing domain-specific knowledge in the prompt design process become important.

Static Prediction of Runtime Errors by Learning to Execute Programs with External Resource Descriptions

1 code implementation7 Mar 2022 David Bieber, Rishab Goel, Daniel Zheng, Hugo Larochelle, Daniel Tarlow

This presents an interesting machine learning challenge: can we predict runtime errors in a "static" setting, where program execution is not possible?

BIG-bench Machine Learning Inductive Bias +1

PLUR: A Unifying, Graph-Based View of Program Learning, Understanding, and Repair

1 code implementation NeurIPS 2021 Zimin Chen, Vincent Hellendoorn, Pascal Lamblin, Petros Maniatis, Pierre-Antoine Manzagol, Daniel Tarlow, Subhodeep Moitra

Machine learning for understanding and editing source code has recently attracted significant interest, with many developments in new models, new code representations, and new tasks. This proliferation can appear disparate and disconnected, making each approach seemingly unique and incompatible, thus obscuring the core machine learning challenges and contributions. In this work, we demonstrate that the landscape can be significantly simplified by taking a general approach of mapping a graph to a sequence of tokens and pointers. Our main result is to show that 16 recently published tasks of different shapes can be cast in this form, based on which a single model architecture achieves near or above state-of-the-art results on nearly all tasks, outperforming custom models like code2seq and alternative generic models like Transformers. This unification further enables multi-task learning and a series of cross-cutting experiments about the importance of different modeling choices for code understanding and repair tasks. The full framework, called PLUR, is easily extensible to more tasks, and will be open-sourced (https://github. com/google-research/plur).

BIG-bench Machine Learning Multi-Task Learning

Learning Generalized Gumbel-max Causal Mechanisms

1 code implementation NeurIPS 2021 Guy Lorberbom, Daniel D. Johnson, Chris J. Maddison, Daniel Tarlow, Tamir Hazan

To perform counterfactual reasoning in Structural Causal Models (SCMs), one needs to know the causal mechanisms, which provide factorizations of conditional distributions into noise sources and deterministic functions mapping realizations of noise to samples.

counterfactual Counterfactual Reasoning

Beyond In-Place Corruption: Insertion and Deletion In Denoising Probabilistic Models

no code implementations ICML Workshop INNF 2021 Daniel D. Johnson, Jacob Austin, Rianne van den Berg, Daniel Tarlow

Denoising diffusion probabilistic models (DDPMs) have shown impressive results on sequence generation by iteratively corrupting each example and then learning to map corrupted versions back to the original.

Denoising

Structured Denoising Diffusion Models in Discrete State-Spaces

3 code implementations NeurIPS 2021 Jacob Austin, Daniel D. Johnson, Jonathan Ho, Daniel Tarlow, Rianne van den Berg

Here, we introduce Discrete Denoising Diffusion Probabilistic Models (D3PMs), diffusion-like generative models for discrete data that generalize the multinomial diffusion model of Hoogeboom et al. 2021, by going beyond corruption processes with uniform transition probabilities.

Denoising Text Generation

Learning to Extend Program Graphs to Work-in-Progress Code

no code implementations28 May 2021 Xuechen Li, Chris J. Maddison, Daniel Tarlow

Source code spends most of its time in a broken or incomplete state during software development.

Code Completion Variable misuse

Learning to Execute Programs with Instruction Pointer Attention Graph Neural Networks

1 code implementation NeurIPS 2020 David Bieber, Charles Sutton, Hugo Larochelle, Daniel Tarlow

More practically, we evaluate these models on the task of learning to execute partial programs, as might arise if using the model as a heuristic function in program synthesis.

Code Completion Learning to Execute +2

On-the-Fly Adaptation of Source Code Models

no code implementations NeurIPS Workshop CAP 2020 Disha Shrivastava, Hugo Larochelle, Daniel Tarlow

The ability to adapt to unseen, local contexts is an important challenge that successful models of source code must overcome.

Learning Graph Structure With A Finite-State Automaton Layer

1 code implementation NeurIPS 2020 Daniel D. Johnson, Hugo Larochelle, Daniel Tarlow

In practice, edges are used both to represent intrinsic structure (e. g., abstract syntax trees of programs) and more abstract relations that aid reasoning for a downstream task (e. g., results of relevant program analyses).

Variable misuse

Software Engineering Event Modeling using Relative Time in Temporal Knowledge Graphs

no code implementations2 Jul 2020 Kian Ahrabian, Daniel Tarlow, Hehuimin Cheng, Jin L. C. Guo

We present a multi-relational temporal Knowledge Graph based on the daily interactions between artifacts in GitHub, one of the largest social coding platforms.

Knowledge Graphs Link Prediction +1

On-the-Fly Adaptation of Source Code Models using Meta-Learning

no code implementations26 Mar 2020 Disha Shrivastava, Hugo Larochelle, Daniel Tarlow

The ability to adapt to unseen, local contexts is an important challenge that successful models of source code must overcome.

Meta-Learning

Learning to Fix Build Errors with Graph2Diff Neural Networks

no code implementations4 Nov 2019 Daniel Tarlow, Subhodeep Moitra, Andrew Rice, Zimin Chen, Pierre-Antoine Manzagol, Charles Sutton, Edward Aftandilian

A diff specifies how to modify the code's abstract syntax tree, represented in the neural network as a sequence of tokens and of pointers to code locations.

Program Repair

Fast Training of Sparse Graph Neural Networks on Dense Hardware

no code implementations27 Jun 2019 Matej Balog, Bart van Merriënboer, Subhodeep Moitra, Yujia Li, Daniel Tarlow

Graph neural networks have become increasingly popular in recent years due to their ability to naturally encode relational input data and their ability to scale to large graphs by operating on a sparse representation of graph adjacency matrices.

Learning Execution through Neural Code Fusion

no code implementations ICLR 2020 Zhan Shi, Kevin Swersky, Daniel Tarlow, Parthasarathy Ranganathan, Milad Hashemi

In this work, we propose a new approach to use GNNs to learn fused representations of general source code and its execution.

Transfer Learning

Direct Policy Gradients: Direct Optimization of Policies in Discrete Action Spaces

no code implementations NeurIPS 2020 Guy Lorberbom, Chris J. Maddison, Nicolas Heess, Tamir Hazan, Daniel Tarlow

A main benefit of DirPG algorithms is that they allow the insertion of domain knowledge in the form of upper bounds on return-to-go at training time, like is used in heuristic search, while still directly computing a policy gradient.

Neural Networks for Modeling Source Code Edits

no code implementations4 Apr 2019 Rui Zhao, David Bieber, Kevin Swersky, Daniel Tarlow

In this work, we instead treat source code as a dynamic object and tackle the problem of modeling the edits that software developers make to source code files.

AMPNet: Asynchronous Model-Parallel Training for Dynamic Neural Networks

1 code implementation ICLR 2018 Alexander L. Gaunt, Matthew A. Johnson, Maik Riechert, Daniel Tarlow, Ryota Tomioka, Dimitrios Vytiniotis, Sam Webster

Through an implementation on multi-core CPUs, we show that AMP training converges to the same accuracy as conventional synchronous training algorithms in a similar number of epochs, but utilizes the available hardware more efficiently even for small minibatch sizes, resulting in significantly shorter overall training times.

Batch Policy Gradient Methods for Improving Neural Conversation Models

no code implementations10 Feb 2017 Kirthevasan Kandasamy, Yoram Bachrach, Ryota Tomioka, Daniel Tarlow, David Carter

We study reinforcement learning of chatbots with recurrent neural network architectures when the rewards are noisy and expensive to obtain.

Chatbot Policy Gradient Methods +2

Differentiable Programs with Neural Libraries

no code implementations ICML 2017 Alexander L. Gaunt, Marc Brockschmidt, Nate Kushman, Daniel Tarlow

We develop a framework for combining differentiable programming languages with neural networks.

Differentiable Functional Program Interpreters

1 code implementation7 Nov 2016 John K. Feser, Marc Brockschmidt, Alexander L. Gaunt, Daniel Tarlow

Recent work on differentiable interpreters relaxes the discrete space of programs into a continuous space so that search over programs can be performed using gradient-based optimization.

Program Synthesis

DeepCoder: Learning to Write Programs

3 code implementations7 Nov 2016 Matej Balog, Alexander L. Gaunt, Marc Brockschmidt, Sebastian Nowozin, Daniel Tarlow

We develop a first line of attack for solving programming competition-style problems from input-output examples using deep learning.

Enumerative Search

TerpreT: A Probabilistic Programming Language for Program Induction

no code implementations15 Aug 2016 Alexander L. Gaunt, Marc Brockschmidt, Rishabh Singh, Nate Kushman, Pushmeet Kohli, Jonathan Taylor, Daniel Tarlow

TerpreT is similar to a probabilistic programming language: a model is composed of a specification of a program representation (declarations of random variables) and an interpreter describing how programs map inputs to outputs (a model connecting unknowns to observations).

BIG-bench Machine Learning Probabilistic Programming +2

Fits Like a Glove: Rapid and Reliable Hand Shape Personalization

no code implementations CVPR 2016 David Joseph Tan, Thomas Cashman, Jonathan Taylor, Andrew Fitzgibbon, Daniel Tarlow, Sameh Khamis, Shahram Izadi, Jamie Shotton

We present a fast, practical method for personalizing a hand shape basis to an individual user's detailed hand shape using only a small set of depth images.

Gated Graph Sequence Neural Networks

13 code implementations17 Nov 2015 Yujia Li, Daniel Tarlow, Marc Brockschmidt, Richard Zemel

Graph-structured data appears frequently in domains including chemistry, natural language semantics, social networks, and knowledge bases.

Drug Discovery Graph Classification +2

Candidate Constrained CRFs for Loss-Aware Structured Prediction

no code implementations10 Dec 2014 Faruk Ahmed, Daniel Tarlow, Dhruv Batra

The result is that we can use loss-aware prediction methodology to improve performance of the highly tuned pipeline system.

Image Segmentation Semantic Segmentation +1

Just-In-Time Learning for Fast and Flexible Inference

no code implementations NeurIPS 2014 S. M. Ali Eslami, Daniel Tarlow, Pushmeet Kohli, John Winn

Much of research in machine learning has centered around the search for inference algorithms that are both general-purpose and efficient.

A* Sampling

no code implementations NeurIPS 2014 Chris J. Maddison, Daniel Tarlow, Tom Minka

The problem of drawing samples from a discrete distribution can be converted into a discrete optimization problem.

Consensus Message Passing for Layered Graphical Models

no code implementations27 Oct 2014 Varun Jampani, S. M. Ali Eslami, Daniel Tarlow, Pushmeet Kohli, John Winn

Generative models provide a powerful framework for probabilistic reasoning.

Empirical Minimum Bayes Risk Prediction: How to Extract an Extra Few % Performance from Vision Models with Just Three More Parameters

no code implementations CVPR 2014 Vittal Premachandran, Daniel Tarlow, Dhruv Batra

When building vision systems that predict structured objects such as image segmentations or human poses, a crucial concern is performance under task-specific evaluation measures (e. g. Jaccard Index or Average Precision).

Structured Generative Models of Natural Source Code

no code implementations2 Jan 2014 Chris J. Maddison, Daniel Tarlow

We study the problem of building generative models of natural source code (NSC); that is, source code written and understood by humans.

Detecting Parameter Symmetries in Probabilistic Models

no code implementations19 Dec 2013 Robert Nishihara, Thomas Minka, Daniel Tarlow

Probabilistic models often have parameters that can be translated, scaled, permuted, or otherwise transformed without changing the model.

Probabilistic Programming

Learning to Pass Expectation Propagation Messages

no code implementations NeurIPS 2013 Nicolas Heess, Daniel Tarlow, John Winn

Expectation Propagation (EP) is a popular approximate posterior inference algorithm that often provides a fast and accurate alternative to sampling-based methods.

Tighter Linear Program Relaxations for High Order Graphical Models

no code implementations26 Sep 2013 Elad Mezuman, Daniel Tarlow, Amir Globerson, Yair Weiss

In this work, we study the LP relaxations that result from enforcing additional consistency constraints between the HOP and the rest of the model.

Vocal Bursts Intensity Prediction

Exploring Compositional High Order Pattern Potentials for Structured Output Learning

no code implementations CVPR 2013 Yujia Li, Daniel Tarlow, Richard Zemel

In this work, we study the learning of a general class of pattern-like high order potential, which we call Compositional High Order Pattern Potentials (CHOPPs).

Vocal Bursts Intensity Prediction

Cardinality Restricted Boltzmann Machines

no code implementations NeurIPS 2012 Kevin Swersky, Ilya Sutskever, Daniel Tarlow, Richard S. Zemel, Ruslan R. Salakhutdinov, Ryan P. Adams

The Restricted Boltzmann Machine (RBM) is a popular density model that is also good for extracting features.

Cannot find the paper you are looking for? You can Submit a new open access paper.