Search Results for author: Katiana Kontolati

Found 7 papers, 4 papers with code

Learning in latent spaces improves the predictive accuracy of deep neural operators

1 code implementation15 Apr 2023 Katiana Kontolati, Somdatta Goswami, George Em Karniadakis, Michael D. Shields

Operator regression provides a powerful means of constructing discretization-invariant emulators for partial-differential equations (PDEs) describing physical systems.

Computational Efficiency

Deep transfer operator learning for partial differential equations under conditional shift

1 code implementation20 Apr 2022 Somdatta Goswami, Katiana Kontolati, Michael D. Shields, George Em Karniadakis

Transfer learning (TL) enables the transfer of knowledge gained in learning to perform one task (source) to a related but different task (target), hence addressing the expense of data acquisition and labeling, potential computational power limitations, and dataset distribution mismatches.

Domain Adaptation Operator learning +4

On the influence of over-parameterization in manifold based surrogates and deep neural operators

1 code implementation9 Mar 2022 Katiana Kontolati, Somdatta Goswami, Michael D. Shields, George Em Karniadakis

In contrast, an even highly over-parameterized DeepONet leads to better generalization for both smooth and non-smooth dynamics.

Operator learning

A survey of unsupervised learning methods for high-dimensional uncertainty quantification in black-box-type problems

no code implementations9 Feb 2022 Katiana Kontolati, Dimitrios Loukrezis, Dimitris G. Giovanis, Lohit Vandanapu, Michael D. Shields

Constructing surrogate models for uncertainty quantification (UQ) on complex partial differential equations (PDEs) having inherently high-dimensional $\mathcal{O}(10^{\ge 2})$ stochastic inputs (e. g., forcing terms, boundary conditions, initial conditions) poses tremendous challenges.

blind source separation Dimensionality Reduction +1

Grassmannian diffusion maps based surrogate modeling via geometric harmonics

no code implementations28 Sep 2021 Ketson R. M. dos Santos, Dimitrios G. Giovanis, Katiana Kontolati, Dimitrios Loukrezis, Michael D. Shields

Using this representation, geometric harmonics, an out-of-sample function extension technique, is employed to create a global map from the space of input parameters to a Grassmannian diffusion manifold.

Uncertainty Quantification

Manifold learning-based polynomial chaos expansions for high-dimensional surrogate models

2 code implementations21 Jul 2021 Katiana Kontolati, Dimitrios Loukrezis, Ketson R. M. dos Santos, Dimitrios G. Giovanis, Michael D. Shields

For this purpose, we employ Grassmannian diffusion maps, a two-step nonlinear dimension reduction technique which allows us to reduce the dimensionality of the data and identify meaningful geometric descriptions in a parsimonious and inexpensive manner.

Dimensionality Reduction Uncertainty Quantification +1

Cannot find the paper you are looking for? You can Submit a new open access paper.