Search Results for author: Konstantinos Spiliopoulos

Found 14 papers, 4 papers with code

Kernel Limit of Recurrent Neural Networks Trained on Ergodic Data Sequences

1 code implementation28 Aug 2023 Samuel Chun-Hei Lam, Justin Sirignano, Konstantinos Spiliopoulos

Mathematical methods are developed to characterize the asymptotics of recurrent neural networks (RNN) as the number of hidden units, data samples in the sequence, hidden state updates, and training steps simultaneously grow to infinity.

Transport map unadjusted Langevin algorithms: learning and discretizing perturbed samplers

no code implementations14 Feb 2023 Benjamin J. Zhang, Youssef M. Marzouk, Konstantinos Spiliopoulos

We show that in continuous time, when a transport map is applied to Langevin dynamics, the result is a Riemannian manifold Langevin dynamics (RMLD) with metric defined by the transport map.

Normalization effects on deep neural networks

1 code implementation2 Sep 2022 Jiahui Yu, Konstantinos Spiliopoulos

A given layer $i$ with $N_{i}$ hidden units is allowed to be normalized by $1/N_{i}^{\gamma_{i}}$ with $\gamma_{i}\in[1/2, 1]$ and we study the effect of the choice of the $\gamma_{i}$ on the statistical behavior of the neural network's output (such as variance) as well as on the test accuracy on the MNIST data set.

Geometry-informed irreversible perturbations for accelerated convergence of Langevin dynamics

no code implementations18 Aug 2021 Benjamin J. Zhang, Youssef M. Marzouk, Konstantinos Spiliopoulos

We introduce a novel geometry-informed irreversible perturbation that accelerates convergence of the Langevin algorithm for Bayesian computation.

PDE-constrained Models with Neural Network Terms: Optimization and Global Convergence

no code implementations18 May 2021 Justin Sirignano, Jonathan MacArt, Konstantinos Spiliopoulos

Recent research has used deep learning to develop partial differential equation (PDE) models in science and engineering.

Normalization effects on shallow neural networks and related asymptotic expansions

1 code implementation20 Nov 2020 Jiahui Yu, Konstantinos Spiliopoulos

In addition, we show that to leading order in $N$, the variance of the neural network's statistical output decays as the implied normalization by the scaling parameter approaches the mean field normalization.

Asymptotics of Reinforcement Learning with Neural Networks

no code implementations13 Nov 2019 Justin Sirignano, Konstantinos Spiliopoulos

In addition, we study the convergence of the limit differential equation to the stationary solution.

Q-Learning reinforcement-learning +1

Scaling Limit of Neural Networks with the Xavier Initialization and Convergence to a Global Minimum

no code implementations9 Jul 2019 Justin Sirignano, Konstantinos Spiliopoulos

We analyze single-layer neural networks with the Xavier initialization in the asymptotic regime of large numbers of hidden units and large numbers of stochastic gradient descent training steps.

Mean Field Analysis of Deep Neural Networks

no code implementations11 Mar 2019 Justin Sirignano, Konstantinos Spiliopoulos

The limit procedure is valid for any number of hidden layers and it naturally also describes the limiting behavior of the training loss.

valid

Information geometry for approximate Bayesian computation

no code implementations5 Dec 2018 Konstantinos Spiliopoulos

We use relative entropy ideas to analyze the behavior of the algorithm as a function of the threshold parameter and of the size of the data.

Mean Field Analysis of Neural Networks: A Central Limit Theorem

no code implementations28 Aug 2018 Justin Sirignano, Konstantinos Spiliopoulos

We rigorously prove a central limit theorem for neural network models with a single hidden layer.

Speech Recognition

Stochastic Gradient Descent in Continuous Time: A Central Limit Theorem

no code implementations11 Oct 2017 Justin Sirignano, Konstantinos Spiliopoulos

Stochastic gradient descent in continuous time (SGDCT) provides a computationally efficient method for the statistical learning of continuous-time models, which are widely used in science, engineering, and finance.

DGM: A deep learning algorithm for solving partial differential equations

8 code implementations24 Aug 2017 Justin Sirignano, Konstantinos Spiliopoulos

The algorithm is tested on a class of high-dimensional free boundary PDEs, which we are able to accurately solve in up to $200$ dimensions.

Stochastic Gradient Descent in Continuous Time

no code implementations17 Nov 2016 Justin Sirignano, Konstantinos Spiliopoulos

Stochastic gradient descent in continuous time (SGDCT) provides a computationally efficient method for the statistical learning of continuous-time models, which are widely used in science, engineering, and finance.

Cannot find the paper you are looking for? You can Submit a new open access paper.