no code implementations • 23 Apr 2024 • Hyeontae Jo, Sung Woong Cho, Hyung Ju Hwang
With RCS data, we found that traditional methods for parameter estimation in differential equations, such as using mean values of time trajectories or Gaussian Process-based trajectory generation, have limitations in estimating the shape of parameter distributions, often leading to a significant loss of data information.
no code implementations • 14 Feb 2024 • Namkyeong Cho, Junseung Ryu, Hyung Ju Hwang
This study investigates the impact of Sobolev Training on operator learning frameworks for improving model performance.
no code implementations • 13 Feb 2024 • Sung Woong Cho, Jae Yong Lee, Hyung Ju Hwang
There has been growing interest in models that learn the operator from the parameters of a partial differential equation (PDE) to the corresponding solutions.
no code implementations • 26 Dec 2023 • Jae Yong Lee, Sung Woong Cho, Hyung Ju Hwang
This study proposes HyperDeepONet, which uses the expressive power of the hypernetwork to enable the learning of a complex operator with a smaller set of parameters.
no code implementations • 5 Jul 2022 • Jae Yong Lee, Juhi Jang, Hyung Ju Hwang
We propose a hybrid framework opPINN: physics-informed neural network (PINN) with operator learning for approximating the solution to the Fokker-Planck-Landau (FPL) equation.
1 code implementation • 29 Apr 2022 • Hwijae Son, Sung Woong Cho, Hyung Ju Hwang
By employing Augmented Lagrangian relaxation, the constrained optimization problem becomes a sequential max-min problem so that the learnable parameters $\lambda$ adaptively balance each loss component.
1 code implementation • 28 Jan 2022 • Jin Young Shin, Jae Yong Lee, Hyung Ju Hwang
We combine the PDIO with the neural operator to develop a \textit{pseudo-differential neural operator} (PDNO) and learn the nonlinear solution operator of PDEs.
no code implementations • 9 Nov 2021 • Rakhoon Hwang, Jae Yong Lee, Jin Young Shin, Hyung Ju Hwang
Once the surrogate model is trained in Phase 1, the optimal control can be inferred in Phase 2 without intensive computations.
no code implementations • 18 Sep 2021 • Min Sue Park, Hyeontae Jo, Haeun Lee, Se Young Jung, Hyung Ju Hwang
The prompt severity assessment model for managing infectious people has been attained through using a nationwide dataset.
no code implementations • 23 Jun 2021 • Hyung Ju Hwang, Hwijae Son
In this paper, we propose a novel conservative formulation for solving kinetic equations via neural networks.
no code implementations • 22 Jan 2021 • Jin Young Shin, Cheolhyeong Kim, Hyung Ju Hwang
In this paper, we claim that active inference can be interpreted using reinforcement learning (RL) algorithms and find a theoretical connection between them.
no code implementations • 21 Jan 2021 • Sung Woong Cho, Hyung Ju Hwang, Hwijae Son
This paper focuses on how to approximate traveling wave solutions for various kinds of partial differential equations via artificial neural networks.
Numerical Analysis Numerical Analysis Analysis of PDEs
no code implementations • 1 Jan 2021 • Hwijae Son, Jin Woo Jang, Woo Jin Han, Hyung Ju Hwang
Inspired by the recent studies that incorporate derivative information for the training of neural networks, we develop a loss function that guides a neural network to reduce the error in the corresponding Sobolev space.
no code implementations • 1 Jan 2021 • Jin Young Shin, Cheolhyeong Kim, Hyung Ju Hwang
In this paper, we claim that active inference can be interpreted using reinforcement learning (RL) algorithms and find a theoretical connection between them.
no code implementations • 28 Sep 2020 • Jae Yong Lee, Jin Woo Jang, Hyung Ju Hwang
The model reduction of a mesoscopic kinetic dynamics to a macroscopic continuum dynamics has been one of the fundamental questions in mathematical physics since Hilbert's time.
no code implementations • 28 Aug 2020 • Taehyun Kim, Hyomin Shin, Hyung Ju Hwang, Seungwon Jeong
Comparing the F1-scores, the features we created outperformed the features used for bot detection on Facebook and Twitter.
no code implementations • 22 Nov 2019 • Hyung Ju Hwang, Jin Woo Jang, Hyeontae Jo, Jae Yong Lee
The issue of the relaxation to equilibrium has been at the core of the kinetic theory of rarefied gas dynamics.
no code implementations • 7 Nov 2019 • Rakhoon Hwang, Hanjin Lee, Hyung Ju Hwang
In this paper, we solve a hierarchical inverse reinforcement learning problem within the options framework, which allows us to utilize intrinsic motivation of the expert demonstrations.
no code implementations • 25 Sep 2019 • Eun saem Lee, Hyung Ju Hwang
The universality of a neural network enables the approximation of any type of continuous functions.
no code implementations • 6 Sep 2019 • Cheolhyeong Kim, Haeseong Moon, Hyung Ju Hwang
However, past GNN algorithms based on 1-hop neighborhood neural message passing are exposed to a risk of loss of information on local structures and relationships.
no code implementations • 27 Jul 2019 • Hyeontae Jo, Hwijae Son, Hyung Ju Hwang, Eunheui Kim
That is, we provide a unified framework of DNN architecture that approximates an analytic solution and its model parameters simultaneously.
no code implementations • ICLR 2019 • Cheolhyeong Kim, Seungtae Park, Hyung Ju Hwang
Wasserstein GAN(WGAN) is a model that minimizes the Wasserstein distance between a data distribution and sample distribution.
no code implementations • 5 Oct 2018 • Cheolhyeong Kim, Seungtae Park, Hyung Ju Hwang
Wasserstein GAN(WGAN) is a model that minimizes the Wasserstein distance between a data distribution and sample distribution.