Search Results for author: Cristiana Diaconu

Found 5 papers, 3 papers with code

On conditional diffusion models for PDE simulations

1 code implementation21 Oct 2024 Aliaksandra Shysheya, Cristiana Diaconu, Federico Bergamin, Paris Perdikaris, José Miguel Hernández-Lobato, Richard E. Turner, Emile Mathieu

Modelling partial differential equations (PDEs) is of crucial importance in science and engineering, and it includes tasks ranging from forecasting to inverse problems, such as data assimilation.

In-Context In-Context Learning with Transformer Neural Processes

no code implementations19 Jun 2024 Matthew Ashman, Cristiana Diaconu, Adrian Weller, Richard E. Turner

Standard NP architectures, such as the convolutional conditional NP (ConvCNP) or the family of transformer neural processes (TNPs), are not capable of in-context in-context learning, as they are only able to condition on a single dataset.

In-Context Learning Meta-Learning

Approximately Equivariant Neural Processes

1 code implementation19 Jun 2024 Matthew Ashman, Cristiana Diaconu, Adrian Weller, Wessel Bruinsma, Richard E. Turner

Our approach is agnostic to both the choice of symmetry group and model architecture, making it widely applicable.

Meta-Learning

Translation Equivariant Transformer Neural Processes

1 code implementation18 Jun 2024 Matthew Ashman, Cristiana Diaconu, Junhyuck Kim, Lakee Sivaraya, Stratis Markou, James Requeima, Wessel P. Bruinsma, Richard E. Turner

Notably, the posterior prediction maps for data that are stationary -- a common assumption in spatio-temporal modelling -- exhibit translation equivariance.

Translation

Cannot find the paper you are looking for? You can Submit a new open access paper.