1 code implementation • 12 Feb 2024 • Biraj Pandey, Bamdad Hosseini, Pau Batlle, Houman Owhadi
This article presents a general framework for the transport of probability measures towards minimum divergence generative modeling and sampling using ordinary differential equations (ODEs) and Reproducing Kernel Hilbert Spaces (RKHSs), inspired by ideas from diffeomorphic matching and image registration.
1 code implementation • 9 Nov 2023 • Bamdad Hosseini, Alexander W. Hsu, Amirhossein Taghvaei
We present a systematic study of conditional triangular transport maps in function spaces from the perspective of optimal transportation and with a view towards amortized Bayesian inference.
no code implementations • 21 Oct 2023 • Mohammad Al-jarrah, Niyizhen Jin, Bamdad Hosseini, Amirhossein Taghvaei
This paper is concerned with the problem of nonlinear filtering, i. e., computing the conditional distribution of the state of a stochastic dynamical system given a history of noisy partial observations.
no code implementations • 8 May 2023 • Pau Batlle, Yifan Chen, Bamdad Hosseini, Houman Owhadi, Andrew M Stuart
We introduce a priori Sobolev-space error estimates for the solution of nonlinear, and possibly parametric, PDEs using Gaussian process and kernel based methods.
1 code implementation • 26 Apr 2023 • Pau Batlle, Matthieu Darcy, Bamdad Hosseini, Houman Owhadi
We present a general kernel-based framework for learning operators between Banach spaces along with a priori error analysis and comprehensive numerical comparisons with popular neural net (NN) approaches such as Deep Operator Net (DeepONet) [Lu et al.] and Fourier Neural Operator (FNO) [Li et al.].
no code implementations • 2 Mar 2023 • Alfredo Garbuno-Inigo, Tapio Helin, Franca Hoffmann, Bamdad Hosseini
In recent years, Bayesian inference in large-scale inverse problems found in science, engineering and machine learning has gained significant attention.
no code implementations • 14 Oct 2022 • Da Long, Nicole Mrvaljevic, Shandian Zhe, Bamdad Hosseini
This article presents a three-step framework for learning and solving partial differential equations (PDEs) using kernel methods.
no code implementations • 22 Mar 2022 • Amirhossein Taghvaei, Bamdad Hosseini
This paper presents a variational representation of the Bayes' law using optimal transportation theory.
2 code implementations • 24 Mar 2021 • Yifan Chen, Bamdad Hosseini, Houman Owhadi, Andrew M Stuart
The main idea of our method is to approximate the solution of a given PDE as the maximum a posteriori (MAP) estimator of a Gaussian process conditioned on solving the PDE at a finite number of collocation points.
no code implementations • 25 Jul 2020 • Andrea L. Bertozzi, Bamdad Hosseini, Hao Li, Kevin Miller, Andrew M. Stuart
Graph-based semi-supervised regression (SSR) is the problem of estimating the value of a function on a weighted graph from its values (labels) on a small subset of the vertices.
1 code implementation • 11 Jun 2020 • Ricardo Baptista, Bamdad Hosseini, Nikola B. Kovachki, Youssef Marzouk
We present a novel framework for conditional sampling of probability measures, using block triangular transport maps.
no code implementations • 7 May 2020 • Kaushik Bhattacharya, Bamdad Hosseini, Nikola B. Kovachki, Andrew M. Stuart
We develop a general framework for data-driven approximation of input-output maps between infinite-dimensional spaces.
no code implementations • 13 Sep 2019 • Franca Hoffmann, Bamdad Hosseini, Assad A. Oberai, Andrew M. Stuart
Graph Laplacians computed from weighted adjacency matrices are widely used to identify geometric structure in data, and clusters in particular; their spectral properties play a central role in a number of unsupervised and semi-supervised learning algorithms.
no code implementations • 18 Jun 2019 • Franca Hoffmann, Bamdad Hosseini, Zhi Ren, Andrew M. Stuart
Graph-based semi-supervised learning is the problem of propagating labels from a small number of labelled data points to a larger set of unlabelled data.
no code implementations • 30 Jan 2019 • Nicolas Garcia Trillos, Franca Hoffmann, Bamdad Hosseini
More precisely, we assume that the data is sampled from a mixture model supported on a manifold $\mathcal{M}$ embedded in $\mathbb{R}^d$, and pick a connectivity length-scale $\varepsilon>0$ to construct a kernelized graph Laplacian.