no code implementations • 12 Jan 2025 • Sanghyun Hong, Fan Wu, Anthony Gruber, Kookjin Lee
By accurately learning underlying dynamics in data in the form of differential equations, NODEs have been widely adopted in various domains, such as healthcare, finance, computer vision, and language modeling.
no code implementations • 9 Oct 2024 • Mingu Kang, Dongseok Lee, Woojin Cho, Jaehyeon Park, Kookjin Lee, Anthony Gruber, Youngjoon Hong, Noseong Park
Large language models (LLMs), like ChatGPT, have shown that even trained with noisy prior data, they can generalize effectively to new tasks through in-context learning (ICL) and pre-training techniques.
no code implementations • 8 Oct 2024 • Jonas A. Actor, Anthony Gruber, Eric C. Cyr, Nathaniel Trask
A machine-learnable variational scheme using Gaussian radial basis functions (GRBFs) is presented and used to approximate linear problems on bounded and unbounded domains.
no code implementations • 2 Sep 2024 • Ian Moore, Christopher Wentland, Anthony Gruber, Irina Tezaur
This paper presents and evaluates an approach for coupling together subdomain-local reduced order models (ROMs) constructed via non-intrusive operator inference (OpInf) with each other and with subdomain-local full order models (FOMs), following a domain decomposition of the spatial geometry on which a given partial differential equation (PDE) is posed.
no code implementations • 25 May 2024 • Anthony Gruber, Kookjin Lee, Haksoo Lim, Noseong Park, Nathaniel Trask
Metriplectic systems are learned from data in a way that scales quadratically in both the size of the state and the rank of the metriplectic data.
1 code implementation • NeurIPS 2023 • Anthony Gruber, Kookjin Lee, Nathaniel Trask
Recent works have shown that physics-inspired architectures allow the training of deep graph neural networks (GNNs) without oversmoothing.
1 code implementation • 13 Apr 2023 • Anthony Gruber, Irina Tezaur
A method for the nonintrusive and structure-preserving model reduction of canonical and noncanonical Hamiltonian systems is presented.
2 code implementations • 2 Dec 2021 • Yuankai Teng, Zhu Wang, Lili Ju, Anthony Gruber, Guannan Zhang
Our method contains two major components: one is the pseudo-reversible neural network (PRNN) module that effectively transforms high-dimensional input variables to low-dimensional active variables, and the other is the synthesized regression module for approximating function values based on the transformed data in the low-dimensional space.
1 code implementation • 5 Oct 2021 • Anthony Gruber, Max Gunzburger, Lili Ju, Zhu Wang
The popularity of deep convolutional autoencoders (CAEs) has engendered new and effective reduced-order models (ROMs) for the simulation of large-scale dynamical systems.
1 code implementation • 29 Apr 2021 • Anthony Gruber, Max Gunzburger, Lili Ju, Yuankai Teng, Zhu Wang
A dimension reduction method based on the "Nonlinear Level set Learning" (NLL) approach is presented for the pointwise prediction of functions which have been sparsely sampled.
no code implementations • 15 Dec 2019 • Anthony Gruber, Magdalena Toda, Hung Tran
This article investigates stationary surfaces with boundaries, which arise as the critical points of functionals dependent on curvature.
Differential Geometry Primary 53A05, Secondary 53A10, 53C40, 53C42