1 code implementation • 1 Feb 2024 • Sifan Wang, Bowen Li, Yuhan Chen, Paris Perdikaris
While physics-informed neural networks (PINNs) have become a popular deep learning framework for tackling forward and inverse problems governed by partial differential equations (PDEs), their performance is known to degrade when larger and deeper neural network architectures are employed.
no code implementations • 24 Aug 2023 • Zhiwei Fang, Sifan Wang, Paris Perdikaris
By reformulating the PDEs into boundary integral equations (BIEs), we can train the operator network solely on the boundary of the domain.
1 code implementation • 16 Aug 2023 • Sifan Wang, Shyam Sankaran, Hanwen Wang, Paris Perdikaris
Physics-informed neural networks (PINNs) have been popularized as a deep learning framework that can seamlessly synthesize observational data and partial differential equation (PDE) constraints.
1 code implementation • 18 May 2023 • Shunyuan Mao, Ruobing Dong, Lu Lu, Kwang Moo Yi, Sifan Wang, Paris Perdikaris
We develop a tool, which we name Protoplanetary Disk Operator Network (PPDONet), that can predict the solution of disk-planet interactions in protoplanetary disks in real-time.
no code implementations • 25 Feb 2023 • Zhiwei Fang, Sifan Wang, Paris Perdikaris
While the popularity of physics-informed neural networks (PINNs) is steadily rising, to this date, PINNs have not been successful in simulating multi-scale and singular perturbation problems.
1 code implementation • 3 Oct 2022 • Sifan Wang, Hanwen Wang, Jacob H. Seidman, Paris Perdikaris
Continuous neural representations have recently emerged as a powerful and flexible alternative to classical discretized representations of signals.
1 code implementation • 5 Jul 2022 • Arka Daw, Jie Bu, Sifan Wang, Paris Perdikaris, Anuj Karpatne
In this paper, we provide a novel perspective of failure modes of PINNs by hypothesizing that training PINNs relies on successful "propagation" of solution from initial and/or boundary condition points to interior points.
3 code implementations • 14 Mar 2022 • Sifan Wang, Shyam Sankaran, Paris Perdikaris
While the popularity of physics-informed neural networks (PINNs) is steadily rising, to this date PINNs have not been successful in simulating dynamical systems whose solution exhibits multi-scale, chaotic or turbulent behavior.
1 code implementation • 25 Oct 2021 • Sifan Wang, Mohamed Aziz Bhouri, Paris Perdikaris
Design and optimal control problems are among the fundamental, ubiquitous tasks we face in science and engineering.
1 code implementation • 4 Oct 2021 • Sifan Wang, Hanwen Wang, Paris Perdikaris
In this work we analyze the training dynamics of deep operator networks (DeepONets) through the lens of Neural Tangent Kernel (NTK) theory, and reveal a bias that favors the approximation of functions with larger magnitudes.
1 code implementation • 9 Jun 2021 • Sifan Wang, Paris Perdikaris
Ordinary and partial differential equations (ODEs/PDEs) play a paramount role in analyzing and simulating complex dynamic processes across all corners of science and engineering.
2 code implementations • 19 Mar 2021 • Sifan Wang, Hanwen Wang, Paris Perdikaris
Deep operator networks (DeepONets) are receiving increased attention thanks to their demonstrated capability to approximate nonlinear operators between infinite-dimensional Banach spaces.
1 code implementation • 18 Dec 2020 • Sifan Wang, Hanwen Wang, Paris Perdikaris
Physics-informed neural networks (PINNs) are demonstrating remarkable promise in integrating physical models with gappy and noisy observational data, but they still struggle in cases where the target functions to be approximated exhibit high-frequency or multi-scale features.
1 code implementation • 28 Jul 2020 • Sifan Wang, Xinling Yu, Paris Perdikaris
In this work, we aim to investigate these questions through the lens of the Neural Tangent Kernel (NTK); a kernel that captures the behavior of fully-connected neural networks in the infinite width limit during training via gradient descent.
1 code implementation • 4 Jun 2020 • Sifan Wang, Paris Perdikaris
Free boundary problems appear naturally in numerous areas of mathematics, science and engineering.
1 code implementation • 13 Jan 2020 • Sifan Wang, Yujun Teng, Paris Perdikaris
The widespread use of neural networks across different scientific domains often involves constraining them to satisfy certain symmetries, conservation laws, or other domain knowledge.
no code implementations • 12 Aug 2019 • Dawei Li, Yan Cao, Guoliang Shi, Xin Cai, Yang Chen, Sifan Wang, Siyuan Yan
The proposed method can also facilitate the automatic traits estimation of each single leaf (such as the leaf area, length, and width), which has potential to become a highly effective tool for plant research and agricultural engineering.
no code implementations • 2 Jul 2019 • Dawei Li, Siyuan Yan, Xin Cai, Yan Cao, Sifan Wang
In this paper, we present an integrated filter which comprises a weighted local guided image filter and a weighted spatiotemporal tree filter.