Search Results for author: Zhiyan Ding

Found 9 papers, 0 papers with code

Convergence of variational Monte Carlo simulation and scale-invariant pre-training

no code implementations21 Mar 2023 Nilin Abrahamsen, Zhiyan Ding, Gil Goldshlager, Lin Lin

We provide theoretical convergence bounds for the variational Monte Carlo (VMC) method as applied to optimize neural network wave functions for the electronic structure problem.

Variational Monte Carlo

On the Global Convergence of Gradient Descent for multi-layer ResNets in the mean-field regime

no code implementations6 Oct 2021 Zhiyan Ding, Shi Chen, Qin Li, Stephen Wright

Finding the optimal configuration of parameters in ResNet is a nonconvex minimization problem, but first-order methods nevertheless find the global optimum in the overparameterized regime.

Overparameterization of deep ResNet: zero loss and mean-field analysis

no code implementations30 May 2021 Zhiyan Ding, Shi Chen, Qin Li, Stephen Wright

Finding parameters in a deep neural network (NN) that fit training data is a nonconvex optimization problem, but a basic first-order optimization method (gradient descent) finds a global optimizer with perfect fit (zero-loss) in many practical situations.

Constrained Ensemble Langevin Monte Carlo

no code implementations8 Feb 2021 Zhiyan Ding, Qin Li

In particular, we find that if one directly surrogates the gradient using the ensemble approximation, the algorithm, termed Ensemble Langevin Monte Carlo, is unstable due to a high variance term.

Random Coordinate Underdamped Langevin Monte Carlo

no code implementations22 Oct 2020 Zhiyan Ding, Qin Li, Jianfeng Lu, Stephen J. Wright

We investigate the computational complexity of RC-ULMC and compare it with the classical ULMC for strongly log-concave probability distributions.

Random Coordinate Langevin Monte Carlo

no code implementations3 Oct 2020 Zhiyan Ding, Qin Li, Jianfeng Lu, Stephen J. Wright

We investigate the total complexity of RC-LMC and compare it with the classical LMC for log-concave probability distributions.

Langevin Monte Carlo: random coordinate descent and variance reduction

no code implementations26 Jul 2020 Zhiyan Ding, Qin Li

However, the method requires the evaluation of a full gradient in each iteration, and for a problem on $\mathbb{R}^d$, this amounts to $d$ times partial derivative evaluations per iteration.

Computational Efficiency

Variance reduction for Random Coordinate Descent-Langevin Monte Carlo

no code implementations NeurIPS 2020 Zhiyan Ding, Qin Li

The high variance induced by the randomness means a larger number of iterations are needed, and this balances out the saving in each iteration.

Error Lower Bounds of Constant Step-size Stochastic Gradient Descent

no code implementations18 Oct 2019 Zhiyan Ding, Yiding Chen, Qin Li, Xiaojin Zhu

To our knowledge, this is the first analysis for SGD error lower bound without the strong convexity assumption.

BIG-bench Machine Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.