no code implementations • 25 Oct 2024 • Xiaoyu Wang, Xuxing Chen, Shiqian Ma, Tong Zhang
This paper focuses on decentralized stochastic bilevel optimization (DSBO) where agents only communicate with their neighbors.
no code implementations • 17 Oct 2024 • Yue Huang, Zhaoxian Wu, Shiqian Ma, Qing Ling
Stochastic approximation (SA) that involves multiple coupled sequences, known as multiple-sequence SA (MSSA), finds diverse applications in the fields of signal processing and machine learning.
1 code implementation • 7 Oct 2024 • Yifan Yang, Hao Ban, Minhui Huang, Shiqian Ma, Kaiyi Ji
To the best of our knowledge, our methods are the first to completely eliminate the need for stepsize tuning, while achieving theoretical guarantees.
no code implementations • 7 Mar 2024 • Ivan Lau, Shiqian Ma, César A. Uribe
Moreover, we propose the decentralized equitable optimal transport (DE-OT) problem.
no code implementations • 15 Nov 2023 • Youran Dong, Shiqian Ma, Junfeng Yang, Chao Yin
Bilevel optimization has gained significant attention in recent years due to its broad applications in machine learning.
no code implementations • 25 Sep 2023 • Jiaxiang Li, Krishnakumar Balasubramanian, Shiqian Ma
We present Zeroth-order Riemannian Averaging Stochastic Approximation (\texttt{Zo-RASA}) algorithms for stochastic optimization on Riemannian manifolds.
1 code implementation • 25 Apr 2023 • Zhong Zheng, Shiqian Ma, Lingzhou Xue
This paper considers the robust phase retrieval problem, which can be cast as a nonsmooth and nonconvex optimization problem.
1 code implementation • 3 Nov 2022 • Jiaxiang Li, Shiqian Ma, Tejes Srivastava
Our algorithm is the first ADMM type algorithm that minimizes a nonsmooth objective over manifold -- a particular nonconvex set.
no code implementations • 23 Oct 2022 • Xuxing Chen, Minhui Huang, Shiqian Ma, Krishnakumar Balasubramanian
Bilevel optimization recently has received tremendous attention due to its great success in solving important machine learning problems like meta learning, reinforcement learning, and hyperparameter optimization.
no code implementations • 12 Jun 2022 • Jiaxiang Li, Shiqian Ma
This paper studies FL over Riemannian manifolds, which finds important applications such as federated PCA and federated kPCA.
no code implementations • 8 Feb 2022 • Minhui Huang, Xuxing Chen, Kaiyi Ji, Shiqian Ma, Lifeng Lai
Moreover, we propose an inexact NEgative-curvature-Originated-from-Noise Algorithm (iNEON), a pure first-order algorithm that can escape saddle point and find local minimum of stochastic bilevel optimization.
no code implementations • 29 Sep 2021 • Minhui Huang, Shiqian Ma, Lifeng Lai
This paper studies the equitable and optimal transport (EOT) problem, which has many applications such as fair division problems and optimal transport with multiple agents etc.
no code implementations • 9 Apr 2021 • Chao Zhang, Xiaojun Chen, Shiqian Ma
In this paper, we propose a Riemannian smoothing steepest descent method to minimize a nonconvex and non-Lipschitz function on submanifolds.
no code implementations • 5 Feb 2021 • Minhui Huang, Shiqian Ma, Lifeng Lai
One of the popular solution methods for this task is to compute the barycenter of the probability measures under the Wasserstein metric.
no code implementations • 9 Dec 2020 • Minhui Huang, Shiqian Ma, Lifeng Lai
We show that the complexity of arithmetic operations for RBCD to obtain an $\epsilon$-stationary point is $O(\epsilon^{-3})$.
no code implementations • 4 Nov 2020 • Yi Zhang, Chengyun Deng, Shiqian Ma, Yongtao Sha, Hui Song
In this paper, a multi-task network is proposed to address both ref-delay estimation and echo cancellation tasks.
no code implementations • 4 Nov 2020 • Chengyun Deng, Shiqian Ma, Yi Zhang, Yongtao Sha, HUI ZHANG, Hui Song, Xiangang Li
dataset confirm the superior performance of the proposed method over the network without IRA in terms of SI-SDR and PESQ improvement.
no code implementations • 18 Aug 2020 • Minhui Huang, Shiqian Ma, Lifeng Lai
This problem aims to decompose a partially observed matrix into the superposition of a low-rank matrix and a sparse matrix, where the sparse matrix captures the grossly corrupted entries of the matrix.
no code implementations • 18 Jul 2020 • Zhongruo Wang, Bingyuan Liu, Shixiang Chen, Shiqian Ma, Lingzhou Xue, Hongyu Zhao
This paper considers a widely adopted model for SSC, which can be formulated as an optimization problem over the Stiefel manifold with nonsmooth and nonconvex objective.
no code implementations • 5 May 2020 • Shixiang Chen, Zengde Deng, Shiqian Ma, Anthony Man-Cho So
Second, we propose a stochastic variant of ManPPA called StManPPA, which is well suited for large-scale computation, and establish its sublinear convergence rate.
no code implementations • 3 May 2020 • Bokun Wang, Shiqian Ma, Lingzhou Xue
However, most of the existing Riemannian stochastic algorithms require the objective function to be differentiable, and they do not apply to the case where the objective function is nonsmooth.
no code implementations • 25 Mar 2020 • Jiaxiang Li, Krishnakumar Balasubramanian, Shiqian Ma
We consider stochastic zeroth-order optimization over Riemannian submanifolds embedded in Euclidean space, where the task is to solve Riemannian optimization problem with only noisy objective function evaluations.
no code implementations • 22 Jan 2020 • Zhongruo Wang, Krishnakumar Balasubramanian, Shiqian Ma, Meisam Razaviyayn
We establish that under the SGC assumption, the complexities of the stochastic algorithms match that of deterministic algorithms.
no code implementations • 15 Jan 2020 • Conghui Tan, Yuqiu Qian, Shiqian Ma, Tong Zhang
Dual averaging-type methods are widely used in industrial machine learning applications due to their ability to promoting solution structure (e. g., sparsity) efficiently.
no code implementations • 27 Mar 2019 • Shixiang Chen, Shiqian Ma, Lingzhou Xue, Hui Zou
Sparse principal component analysis (PCA) and sparse canonical correlation analysis (CCA) are two essential techniques from high-dimensional statistics and machine learning for analyzing large-scale data.
no code implementations • NeurIPS 2018 • Conghui Tan, Tong Zhang, Shiqian Ma, Ji Liu
Regularized empirical risk minimization problem with linear predictor appears frequently in machine learning.
no code implementations • NeurIPS 2018 • Conghui Tan, Tong Zhang, Shiqian Ma, Ji Liu
Regularized empirical risk minimization problem with linear predictor appears frequently in machine learning.
no code implementations • 2 Nov 2018 • Shixiang Chen, Shiqian Ma, Anthony Man-Cho So, Tong Zhang
We prove that the proposed method globally converges to a stationary point.
no code implementations • 9 Jun 2018 • Shiqian Ma, Necdet Serhat Aybat
Robust PCA has drawn significant attention in the last decade due to its success in numerous application domains, ranging from bio-informatics, statistics, and machine learning to image and video processing in computer vision.
1 code implementation • 31 May 2018 • Tianyi Lin, Shiqian Ma, Yinyu Ye, Shuzhong Zhang
Due its connection to Newton's method, IPM is often classified as second-order method -- a genre that is attached with stability and accuracy at the expense of scalability.
Optimization and Control
no code implementations • 6 Feb 2018 • Jason Causey, Junyu Zhang, Shiqian Ma, Bo Jiang, Jake Qualls, David G. Politte, Fred Prior, Shuzhong Zhang, Xiuzhen Huang
Here we present NoduleX, a systematic approach to predict lung nodule malignancy from CT data, based on deep learning convolutional neural networks (CNN).
no code implementations • 5 Oct 2017 • Junyu Zhang, Shiqian Ma, Shuzhong Zhang
For prohibitively large-size tensor or machine learning models, we present a sampling-based stochastic algorithm with the same iteration complexity bound in expectation.
no code implementations • ICML 2017 • Li Shen, Wei Liu, Ganzhao Yuan, Shiqian Ma
In addition, we develop a new technique to establish the global convergence of the GSOS algorithm.
no code implementations • NeurIPS 2017 • Shixiang Chen, Shiqian Ma, Wei Liu
In this paper, we extend the geometric descent method recently proposed by Bubeck, Lee and Singh to tackle nonsmooth and strongly convex composite problems.
no code implementations • 5 Jul 2016 • Xiao Wang, Shiqian Ma, Donald Goldfarb, Wei Liu
In this paper we study stochastic quasi-Newton methods for nonconvex stochastic optimization, where we assume that noisy information about the gradients of the objective function is available via a stochastic first-order oracle (SFO).
no code implementations • NeurIPS 2016 • Conghui Tan, Shiqian Ma, Yu-Hong Dai, Yuqiu Qian
One of the major issues in stochastic gradient descent (SGD) methods is how to choose an appropriate step size while running the algorithm.
no code implementations • 9 May 2016 • Bo Jiang, Tianyi Lin, Shiqian Ma, Shuzhong Zhang
In particular, we consider in this paper some constrained nonconvex optimization models in block decision variables, with or without coupled affine constraints.
no code implementations • 16 May 2015 • Tianyi Lin, Shiqian Ma, Shuzhong Zhang
The alternating direction method of multipliers (ADMM) has been successfully applied to solve structured convex optimization problems due to its superior practical performance.
no code implementations • 26 Sep 2013 • Necdet Serhat Aybat, Donald Goldfarb, Shiqian Ma
Moreover, if the observed data matrix has also been corrupted by a dense noise matrix in addition to gross sparse error, then the stable principal component pursuit (SPCP) problem is solved to recover the low-rank matrix.
Optimization and Control
no code implementations • 27 Jan 2013 • Tianyi Lin, Shiqian Ma, Shuzhong Zhang
The classical alternating direction type methods usually assume that the two convex functions have relatively easy proximal mappings.
no code implementations • NeurIPS 2010 • Katya Scheinberg, Shiqian Ma, Donald Goldfarb
Gaussian graphical models are of great interest in statistical learning.
no code implementations • 23 Dec 2009 • Donald Goldfarb, Shiqian Ma, Katya Scheinberg
We present in this paper first-order alternating linearization algorithms based on an alternating direction augmented Lagrangian approach for minimizing the sum of two convex functions.
1 code implementation • 11 May 2009 • Shiqian Ma, Donald Goldfarb, Lifeng Chen
The tightest convex relaxation of this problem is the linearly constrained nuclear norm minimization.
Optimization and Control Information Theory Information Theory