no code implementations • 20 Oct 2022 • Xinran Zhu, Leo Huang, Cameron Ibrahim, Eric Hans Lee, David Bindel
The Bayesian transformed Gaussian process (BTG) model, proposed by Kedem and Oliviera, is a fully Bayesian counterpart to the warped Gaussian process (WGP) and marginalizes out a joint prior over input warping and kernel hyperparameters.
no code implementations • NeurIPS 2021 • Misha Padidar, Xinran Zhu, Leo Huang, Jacob R. Gardner, David Bindel
We demonstrate the full scalability of our approach on a variety of tasks, ranging from a high dimensional stellarator fusion regression task to training graph convolutional neural networks on Pubmed using Bayesian optimization.
no code implementations • 21 Oct 2020 • Leo Huang, Andrew Graven, David Bindel
A fundamental problem on graph-structured data is that of quantifying similarity between graphs.
3 code implementations • NeurIPS 2020 • Aaron Lou, Derek Lim, Isay Katsman, Leo Huang, Qingxuan Jiang, Ser-Nam Lim, Christopher De Sa
To better conform to data geometry, recent deep generative modelling techniques adapt Euclidean constructions to non-Euclidean spaces.