Search Results for author: Xi-Lin Li

Found 9 papers, 7 papers with code

Stochastic Hessian Fittings with Lie Groups

no code implementations19 Feb 2024 Xi-Lin Li

This paper studies the fitting of Hessian or its inverse for stochastic optimizations using a Hessian fitting criterion from the preconditioned stochastic gradient descent (PSGD) method, which is intimately related to many commonly used second order and adaptive gradient optimizers, e. g., BFGS, Gaussian-Newton and natural gradient descent, AdaGrad, etc.

Curvature-Informed SGD via General Purpose Lie-Group Preconditioners

1 code implementation7 Feb 2024 Omead Pooladzandi, Xi-Lin Li

We present a novel approach to accelerate stochastic gradient descent (SGD) by utilizing curvature information obtained from Hessian-vector products or finite differences of parameters and gradients, similar to the BFGS algorithm.

Independent Vector Analysis with Deep Neural Network Source Priors

1 code implementation23 Aug 2020 Xi-Lin Li

This paper studies the density priors for independent vector analysis (IVA) with convolutive speech mixture separation as the exemplary application.

Speech Separation

A Triangular Network For Density Estimation

1 code implementation30 Apr 2020 Xi-Lin Li

We report a triangular neural network implementation of neural autoregressive flow (NAF).

Density Estimation

A Multiclass Multiple Instance Learning Method with Exact Likelihood

1 code implementation29 Nov 2018 Xi-Lin Li

We study a multiclass multiple instance learning (MIL) problem where the labels only suggest whether any instance of a class exists or does not exist in a training sample or example.

General Classification Multiple Instance Learning +2

Preconditioner on Matrix Lie Group for SGD

2 code implementations ICLR 2019 Xi-Lin Li

We study two types of preconditioners and preconditioned stochastic gradient descent (SGD) methods in a unified framework.

Vocal Bursts Type Prediction

Online Second Order Methods for Non-Convex Stochastic Optimizations

1 code implementation26 Mar 2018 Xi-Lin Li

This paper proposes a family of online second order methods for possibly non-convex stochastic optimizations based on the theory of preconditioned stochastic gradient descent (PSGD), which can be regarded as an enhance stochastic Newton method with the ability to handle gradient noise and non-convexity simultaneously.

Second-order methods

Recurrent neural network training with preconditioned stochastic gradient descent

no code implementations14 Jun 2016 Xi-Lin Li

This paper studies the performance of a recently proposed preconditioned stochastic gradient descent (PSGD) algorithm on recurrent neural network (RNN) training.

Handwritten Digit Recognition

Preconditioned Stochastic Gradient Descent

2 code implementations14 Dec 2015 Xi-Lin Li

When stochastic gradient is used, it can naturally damp the gradient noise to stabilize SGD.

2D Human Pose Estimation

Cannot find the paper you are looking for? You can Submit a new open access paper.