# Gaussian Processes

439 papers with code • 0 benchmarks • 4 datasets

**Gaussian Processes** is a powerful framework for several machine learning tasks such as regression, classification and inference. Given a finite set of input output training data that is generated out of a fixed (but possibly unknown) function, the framework models the unknown function as a stochastic process such that the training outputs are a finite number of jointly Gaussian random variables, whose properties can then be used to infer the statistics (the mean and variance) of the function at test values of input.

## Benchmarks

These leaderboards are used to track progress in Gaussian Processes
## Libraries

Use these libraries to find Gaussian Processes models and implementations## Most implemented papers

# Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning

In comparison, Bayesian models offer a mathematically grounded framework to reason about model uncertainty, but usually come with a prohibitive computational cost.

# Conditional Neural Processes

Deep neural networks excel at function approximation, yet they are typically trained from scratch for each new function.

# Gaussian Processes for Big Data

We introduce stochastic variational inference for Gaussian process models.

# Deep Neural Networks as Gaussian Processes

As such, previous work has not identified that these kernels can be used as covariance functions for GPs and allow fully Bayesian prediction with a deep neural network.

# Doubly Stochastic Variational Inference for Deep Gaussian Processes

Existing approaches to inference in DGP models assume approximate posteriors that force independence between the layers, and do not work well in practice.

# Neural Tangent Kernel: Convergence and Generalization in Neural Networks

While the NTK is random at initialization and varies during training, in the infinite-width limit it converges to an explicit limiting kernel and it stays constant during training.

# Adversarial Robustness Toolbox v1.0.0

Defending Machine Learning models involves certifying and verifying model robustness and model hardening with approaches such as pre-processing inputs, augmenting training data with adversarial samples, and leveraging runtime detection methods to flag any inputs that might have been modified by an adversary.

# Efficiently Sampling Functions from Gaussian Process Posteriors

Gaussian processes are the gold standard for many real-world modeling problems, especially in cases where a model's success hinges upon its ability to faithfully represent predictive uncertainty.

# Scalable Bayesian Optimization Using Deep Neural Networks

Bayesian optimization is an effective methodology for the global optimization of functions with expensive evaluations.

# Deep Kernel Learning

We introduce scalable deep kernels, which combine the structural properties of deep learning architectures with the non-parametric flexibility of kernel methods.