no code implementations • 23 Jan 2023 • Yiling Luo, Yiling Xie, Xiaoming Huo
To compare, we prove that the computational complexity of the Stochastic Sinkhorn algorithm is $\widetilde{{O}}({n^2}/{\epsilon^2})$, which is slower than our accelerated primal-dual stochastic mirror algorithm.
no code implementations • 2 Dec 2022 • Yiling Luo, Xiaoming Huo, Yajun Mei
Our second estimator is a Hessian-free estimator that overcomes the aforementioned limitation.
no code implementations • 29 Oct 2022 • Yiling Xie, Yiling Luo, Xiaoming Huo
Computing the empirical Wasserstein distance in the independence test requires solving this special type of OT problem, where $m=n^2$.
no code implementations • 29 Apr 2022 • Yiling Luo, Xiaoming Huo, Yajun Mei
In addition, the Gradient Descent (GD) with a moderate or small step-size converges along the direction that corresponds to the smallest eigenvalue.
no code implementations • 29 Apr 2022 • Yiling Luo, Xiaoming Huo, Yajun Mei
On the other hand, algorithms such as gradient descent and stochastic gradient descent have the implicit regularization property that leads to better performance in terms of the generalization errors.
1 code implementation • 2 Mar 2022 • Yiling Xie, Yiling Luo, Xiaoming Huo
A primal-dual accelerated stochastic gradient descent with variance reduction algorithm (PDASGD) is proposed to solve linear-constrained optimization problems.
no code implementations • 29 Sep 2021 • Yiling Luo, Xiaoming Huo, Yajun Mei
This paper studies the Stochastic Gradient Descent (SGD) algorithm in kernel regression.