no code implementations • 15 Sep 2024 • Yuxin Dong, Jianhua Yao, Jiajing Wang, Yingbin Liang, ShuHan Liao, Minheng Xiao
Financial fraud refers to the act of obtaining financial benefits through dishonest means.
no code implementations • 5 Sep 2024 • Peizhong Ju, Haibo Yang, Jia Liu, Yingbin Liang, Ness Shroff
This motivates us to investigate a fundamental question in FL: Can we quantify the impact of data heterogeneity and local updates on the generalization performance for FL as the learning process evolves?
no code implementations • 19 Aug 2024 • Tong Yang, Yu Huang, Yingbin Liang, Yuejie Chi
The contextual generalization here can be attained via learning the template function for each task in-context, where all template functions lie in a linear space with $m$ basis functions.
no code implementations • 8 Aug 2024 • Yingbin Liang, Xiqing Liu, Haohao Xia, Yiru Cang, Zitao Zheng, Yuanfang Yang
In this paper, Pro-HRnet-CNN, an innovative model combining HRNet and void-convolution techniques, is proposed for disease prediction under lung imaging.
no code implementations • 7 Aug 2024 • Yihao Zhong, Yijing Wei, Yingbin Liang, Xiqing Liu, Rongwei Ji, Yiru Cang
In this paper, an image recognition algorithm based on the combination of deep learning and generative adversarial network (GAN) is studied, and compared with traditional image recognition methods.
no code implementations • 24 Jun 2024 • Hongbo Li, Sen Lin, Lingjie Duan, Yingbin Liang, Ness B. Shroff
Continual learning (CL) has garnered significant attention because of its ability to adapt to new tasks that arrive over time.
no code implementations • 13 Jun 2024 • Jinyin Wang, Haijing Zhang, Yihao Zhong, Yingbin Liang, Rongwei Ji, Yiru Cang
Experiments show that compared with existing image-text matching models, the optimized new model has significantly improved performance on a series of benchmark data sets.
no code implementations • 4 Mar 2024 • Yu Huang, Zixin Wen, Yuejie Chi, Yingbin Liang
Masked reconstruction, which predicts randomly masked patches from unmasked ones, has emerged as an important approach in self-supervised pretraining.
1 code implementation • 22 Feb 2024 • Xuxi Chen, Zhendong Wang, Daouda Sow, Junjie Yang, Tianlong Chen, Yingbin Liang, Mingyuan Zhou, Zhangyang Wang
Our study starts from an empirical strategy for the light continual training of LLMs using their original pre-training data sets, with a specific focus on selective retention of samples that incur moderately high losses.
no code implementations • 21 Feb 2024 • Yuchen Liang, Peizhong Ju, Yingbin Liang, Ness Shroff
In this paper, we establish the convergence guarantee for substantially larger classes of distributions under DT diffusion processes and further improve the convergence rate for distributions with bounded support.
no code implementations • 5 Feb 2024 • Junze Deng, Yuan Cheng, Shaofeng Zou, Yingbin Liang
Our result for the second model is the first-known result for such a type of function approximation models.
1 code implementation • 3 Dec 2023 • Junjie Yang, Jinze Zhao, Peihao Wang, Zhangyang Wang, Yingbin Liang
However, vanilla ControlNet generally requires extensive training of around 5000 steps to achieve a desirable control for a single task.
1 code implementation • 3 Dec 2023 • Junjie Yang, Tianlong Chen, Xuxi Chen, Zhangyang Wang, Yingbin Liang
Based on that, we further propose a new raw gradient descent (RGD) algorithm that eliminates the use of sign.
no code implementations • 20 Oct 2023 • Ruiquan Huang, Yuan Cheng, Jing Yang, Vincent Tan, Yingbin Liang
To this end, we posit a joint model class for tasks and use the notion of $\eta$-bracketing number to quantify its complexity; this number also serves as a general metric to capture the similarity of tasks and thus determines the benefit of multi-task over single-task RL.
no code implementations • 8 Oct 2023 • Yu Huang, Yuan Cheng, Yingbin Liang
For data with balanced features, we establish the finite-time convergence guarantee with near-zero prediction error by navigating our analysis over two phases of the training dynamics of the attention map.
no code implementations • 17 Aug 2023 • Songtao Feng, Ming Yin, Yu-Xiang Wang, Jing Yang, Yingbin Liang
In this work, we propose a model-free stage-based Q-learning algorithm and show that it achieves the same sample complexity as the best model-based algorithm, and hence for the first time demonstrate that model-free algorithms can enjoy the same optimality in the $H$ dependence as model-based algorithms.
no code implementations • 1 Aug 2023 • Daouda Sow, Sen Lin, Zhangyang Wang, Yingbin Liang
Experiments on standard classification datasets demonstrate that our proposed approach outperforms related state-of-the-art baseline methods in terms of average robust performance, and at the same time improves the robustness against attacks on the weakest data points.
no code implementations • 1 Jul 2023 • Ruiquan Huang, Yingbin Liang, Jing Yang
The general sequential decision-making problem, which includes Markov decision processes (MDPs) and partially observable MDPs (POMDPs) as special cases, aims at maximizing a cumulative reward by making a sequence of decisions based on a history of observations and actions over time.
no code implementations • 14 Jun 2023 • Ming Shi, Yingbin Liang, Ness Shroff
However, existing theoretical results have shown that learning in POMDPs is intractable in the worst case, where the main challenge lies in the lack of latent state information.
no code implementations • 8 Jun 2023 • Peizhong Ju, Sen Lin, Mark S. Squillante, Yingbin Liang, Ness B. Shroff
For example, when the total number of features in the source task's learning model is fixed, we show that it is more advantageous to allocate a greater number of redundant features to the task-specific part rather than the common part.
no code implementations • 1 Jun 2023 • Songtao Feng, Ming Yin, Ruiquan Huang, Yu-Xiang Wang, Jing Yang, Yingbin Liang
To the best of our knowledge, this is the first dynamic regret analysis in non-stationary MDPs with general function approximation.
no code implementations • 9 Apr 2023 • Peizhong Ju, Yingbin Liang, Ness B. Shroff
However, due to the uniqueness of meta-learning such as task-specific gradient descent inner training and the diversity/fluctuation of the ground-truth signals among training tasks, we find new and interesting properties that do not exist in single-task linear regression.
no code implementations • 20 Mar 2023 • Yuan Cheng, Ruiquan Huang, Jing Yang, Yingbin Liang
In this work, we first provide the first known sample complexity lower bound that holds for any algorithm under low-rank MDPs.
1 code implementation • 28 Feb 2023 • Junjie Yang, Xuxi Chen, Tianlong Chen, Zhangyang Wang, Yingbin Liang
This data-driven procedure yields L2O that can efficiently solve problems similar to those seen in training, that is, drawn from the same ``task distribution".
1 code implementation • 22 Feb 2023 • Junjie Yang, Tianlong Chen, Mingkang Zhu, Fengxiang He, DaCheng Tao, Yingbin Liang, Zhangyang Wang
While the optimizer generalization has been recently studied, the optimizee generalization (or learning to generalize) has not been rigorously studied in the L2O context, which is the aim of this paper.
no code implementations • 12 Feb 2023 • Sen Lin, Peizhong Ju, Yingbin Liang, Ness Shroff
In particular, there is a lack of understanding on what factors are important and how they affect "catastrophic forgetting" and generalization performance.
no code implementations • 8 Feb 2023 • Ming Shi, Yingbin Liang, Ness Shroff
In many applications of Reinforcement Learning (RL), it is critically important that the algorithm performs safely, such that instantaneous hard constraints are satisfied at each step, and unsafe states and actions are avoided.
no code implementations • 8 Feb 2023 • Ming Shi, Yingbin Liang, Ness Shroff
Our lower bound indicates that, due to the fundamental challenge of switching costs in adversarial RL, the best achieved regret (whose dependency on $T$ is $\tilde{O}(\sqrt{T})$) in static RL with switching costs (as well as adversarial RL without switching costs) is no longer achievable.
no code implementations • 2 Feb 2023 • Daouda Sow, Sen Lin, Yingbin Liang, Junshan Zhang
More specifically, we first propose two simple but effective detection mechanisms of task switches and distribution shift based on empirical observations, which serve as a key building block for more elegant online model updates in our algorithm: the task switch detection mechanism allows reusing of the best model available for the current task at hand, and the distribution shift detection mechanism differentiates the meta model update in order to preserve the knowledge for in-distribution tasks and quickly learn the new knowledge for out-of-distribution tasks.
no code implementations • 1 Jan 2023 • Hongru Yang, Ziyu Jiang, Ruizhe Zhang, Zhangyang Wang, Yingbin Liang
This work studies training one-hidden-layer overparameterized ReLU networks via gradient descent in the neural tangent kernel (NTK) regime, where the networks' biases are initialized to some constant rather than zero.
no code implementations • 1 Jan 2023 • Hongru Yang, Yingbin Liang, Xiaojie Guo, Lingfei Wu, Zhangyang Wang
It is shown that as long as the pruning fraction is below a certain threshold, gradient descent can drive the training loss toward zero and the network exhibits good generalization performance.
no code implementations • 18 Aug 2022 • Xuyang Chen, Jingliang Duan, Yingbin Liang, Lin Zhao
To our knowledge, this is the first finite-time convergence analysis for the single sample two-timescale AC for solving LQR with global optimality.
no code implementations • 28 Jun 2022 • Ruiquan Huang, Jing Yang, Yingbin Liang
In particular, we consider the scenario where a safe baseline policy is known beforehand, and propose a unified Safe reWard-frEe ExploraTion (SWEET) framework.
no code implementations • 18 Jun 2022 • Yu Huang, Yingbin Liang, Longbo Huang
Despite the superior empirical success of deep meta-learning, theoretical understanding of overparameterized meta-learning is still limited.
no code implementations • 13 Jun 2022 • Tengyu Xu, Yue Wang, Shaofeng Zou, Yingbin Liang
The remarkable success of reinforcement learning (RL) heavily relies on observing the reward of every visited state-action pair.
no code implementations • 13 Jun 2022 • Yuan Cheng, Songtao Feng, Jing Yang, Hong Zhang, Yingbin Liang
To the best of our knowledge, this is the first theoretical study that characterizes the benefit of representation learning in exploration-based reward-free multitask RL for both upstream and downstream tasks.
no code implementations • 27 May 2022 • Kaiyi Ji, Mingrui Liu, Yingbin Liang, Lei Ying
Existing studies in the literature cover only some of those implementation choices, and the complexity bounds available are not refined enough to enable rigorous comparison among different implementations.
no code implementations • 31 Mar 2022 • Shaocong Ma, Ziyi Chen, Yi Zhou, Kaiyi Ji, Yingbin Liang
Moreover, we show that online SGD with mini-batch sampling can further substantially improve the sample complexity over online SGD with periodic data-subsampling over highly dependent data.
no code implementations • 1 Mar 2022 • Daouda Sow, Kaiyi Ji, Ziwei Guan, Yingbin Liang
Existing algorithms designed for such a problem were applicable to restricted situations and do not come with a full guarantee of convergence.
no code implementations • ICLR 2022 • Sen Lin, Jialin Wan, Tengyu Xu, Yingbin Liang, Junshan Zhang
In particular, we devise a new meta-Regularized model-based Actor-Critic (RAC) method for within-task policy optimization, as a key building block of MerPO, using conservative policy evaluation and regularized policy improvement; and the intrinsic tradeoff therein is achieved via striking the right balance between two regularizers, one based on the behavior policy and the other on the meta-policy.
no code implementations • NeurIPS 2021 • Lin Zhao, Huaqing Xiong, Yingbin Liang
This paper tackles the more challenging case of a constant learning rate, and develops new analytical tools that improve the existing convergence rate by orders of magnitude.
no code implementations • 20 Oct 2021 • Tianjiao Li, Ziwei Guan, Shaofeng Zou, Tengyu Xu, Yingbin Liang, Guanghui Lan
Despite the challenge of the nonconcave objective subject to nonconcave constraints, the proposed approach is shown to converge to the global optimum with a complexity of $\tilde{\mathcal O}(1/\epsilon)$ in terms of the optimality gap and the constraint violation, which improves the complexity of the existing primal-dual approach by a factor of $\mathcal O(1/\epsilon)$ \citep{ding2020natural, paternain2019constrained}.
no code implementations • ICLR 2022 • Ziwei Guan, Tengyu Xu, Yingbin Liang
Although ETD has been shown to converge asymptotically to a desirable value function, it is well-known that ETD often encounters a large variance so that its sample complexity can increase exponentially fast with the number of iterations.
1 code implementation • 13 Oct 2021 • Daouda Sow, Kaiyi Ji, Yingbin Liang
Bilevel optimization has arisen as a powerful tool in modern machine learning.
no code implementations • 29 Sep 2021 • Junjie Yang, Tianlong Chen, Mingkang Zhu, Fengxiang He, DaCheng Tao, Yingbin Liang, Zhangyang Wang
Learning to optimize (L2O) has gained increasing popularity in various optimization tasks, since classical optimizers usually require laborious, problem-specific design and hyperparameter tuning.
no code implementations • 29 Sep 2021 • Shaocong Ma, Ziyi Chen, Yi Zhou, Kaiyi Ji, Yingbin Liang
Specifically, with a $\phi$-mixing model that captures both exponential and polynomial decay of the data dependence over time, we show that SGD with periodic data-subsampling achieves an improved sample complexity over the standard SGD in the full spectrum of the $\phi$-mixing data dependence.
no code implementations • 29 Sep 2021 • Daouda Sow, Kaiyi Ji, Yingbin Liang
Bilevel optimization (BO) has arisen as a powerful tool for solving many modern machine learning problems.
no code implementations • 6 Jul 2021 • Tengyu Xu, Zhuoran Yang, Zhaoran Wang, Yingbin Liang
We further show that unlike GTD, the learned GVFs by GenTD are guaranteed to converge to the ground truth GVFs as long as the function approximation power is sufficiently large.
1 code implementation • NeurIPS 2021 • Junjie Yang, Kaiyi Ji, Yingbin Liang
Bilevel optimization has been widely applied in many important machine learning applications such as hyperparameter optimization and meta-learning.
no code implementations • 23 Feb 2021 • Tengyu Xu, Zhuoran Yang, Zhaoran Wang, Yingbin Liang
We also show that the overall convergence of DR-Off-PAC is doubly robust to the approximation errors that depend only on the expressive power of approximation functions.
no code implementations • ICLR 2021 • Ziyi Chen, Yi Zhou, Tengyu Xu, Yingbin Liang
By leveraging this Lyapunov function and the K{\L} geometry that parameterizes the local geometries of general nonconvex functions, we formally establish the variable convergence of proximal-GDA to a critical point $x^*$, i. e., $x_t\to x^*, y_t\to y^*(x^*)$.
no code implementations • 7 Feb 2021 • Kaiyi Ji, Yingbin Liang
Bilevel optimization has recently attracted growing interests due to its wide applications in modern machine learning problems.
no code implementations • 1 Jan 2021 • Lin Zhao, Huaqing Xiong, Yingbin Liang, Wei zhang
Double Q-learning (Hasselt 2010) has gained significant success in practice due to its effectiveness in overcoming the overestimation issue of Q-learning.
no code implementations • 11 Nov 2020 • Tengyu Xu, Yingbin Liang, Guanghui Lan
To demonstrate the theoretical performance of CRPO, we adopt natural policy gradient (NPG) for each policy update step and show that CRPO achieves an $\mathcal{O}(1/\sqrt{T})$ convergence rate to the global optimal policy in the constrained policy set and an $\mathcal{O}(1/\sqrt{T})$ error bound on constraint satisfaction.
no code implementations • 10 Nov 2020 • Tengyu Xu, Yingbin Liang
For linear TDC, we provide a novel non-asymptotic analysis and show that it attains an $\epsilon$-accurate solution with the optimal sample complexity of $\mathcal{O}(\epsilon^{-1}\log(1/\epsilon))$ under a constant stepsize.
2 code implementations • 15 Oct 2020 • Kaiyi Ji, Junjie Yang, Yingbin Liang
For the AID-based method, we orderwisely improve the previous convergence rate analysis due to a more practical parameter selection as well as a warm start strategy, and for the ITD-based method we establish the first theoretical convergence rate.
no code implementations • NeurIPS 2020 • Huaqing Xiong, Lin Zhao, Yingbin Liang, Wei zhang
Although Q-learning is one of the most successful algorithms for finding the best action-value function (and thus the optimal policy) in reinforcement learning, its implementation often suffers from large overestimation of Q-function values incurred by random sampling.
no code implementations • 28 Sep 2020 • Tengyu Xu, Zhe Wang, Yingbin Liang, H. Vincent Poor
Specifically, a novel variance reduction algorithm SREDA was proposed recently by (Luo et al. 2020) to solve such a problem, and was shown to achieve the optimal complexity dependence on the required accuracy level $\epsilon$.
no code implementations • 28 Sep 2020 • Tengyu Xu, Yingbin Liang, Guanghui Lan
To demonstrate the theoretical performance of CRPO, we adopt natural policy gradient (NPG) for each policy update step and show that CRPO achieves an $\mathcal{O}(1/\sqrt{T})$ convergence rate to the global optimal policy in the constrained policy set and an $\mathcal{O}(1/\sqrt{T})$ error bound on constraint satisfaction.
no code implementations • 28 Sep 2020 • Kaiyi Ji, Junjie Yang, Yingbin Liang
For the AID-based method, we orderwisely improve the previous finite-time convergence analysis due to a more practical parameter selection as well as a warm start strategy, and for the ITD-based method we establish the first theoretical convergence rate.
no code implementations • 9 Aug 2020 • Zhe Wang, Yingbin Liang, Pengsheng Ji
Community detection in large social networks is affected by degree heterogeneity of nodes.
no code implementations • 30 Jul 2020 • Bowen Weng, Huaqing Xiong, Lin Zhao, Yingbin Liang, Wei zhang
For the infinite state-action space case, we establish the convergence guarantee for MomentumQ with linear function approximations and Markovian sampling.
no code implementations • 15 Jul 2020 • Bowen Weng, Huaqing Xiong, Yingbin Liang, Wei zhang
In this paper, we first characterize the convergence rate for Q-AMSGrad, which is the Q-learning algorithm with AMSGrad update (a commonly adopted alternative of Adam for theoretical analysis).
no code implementations • 24 Jun 2020 • Ziwei Guan, Tengyu Xu, Yingbin Liang
Generative adversarial imitation learning (GAIL) is a popular inverse reinforcement learning approach for jointly optimizing policy and reward from expert trajectories.
no code implementations • 16 Jun 2020 • Tengyu Xu, Zhe Wang, Yingbin Liang, H. Vincent Poor
In this paper, we focus on such a gradient-free setting, and consider the nonconvex-strongly-concave minimax stochastic optimization problem.
no code implementations • NeurIPS 2020 • Kaiyi Ji, Jason D. Lee, Yingbin Liang, H. Vincent Poor
Although model-agnostic meta-learning (MAML) is a very successful algorithm in meta-learning practice, it can have high computational cost because it updates all model parameters over both the inner loop of task-specific adaptation and the outer-loop of meta initialization training.
no code implementations • 7 May 2020 • Tengyu Xu, Zhe Wang, Yingbin Liang
In the first nested-loop design, actor's one update of policy is followed by an entire loop of critic's updates of the value function, and the finite-sample analysis of such AC and NAC algorithms have been recently well established.
no code implementations • NeurIPS 2020 • Tengyu Xu, Zhe Wang, Yingbin Liang
We show that the overall sample complexity for a mini-batch AC to attain an $\epsilon$-accurate stationary point improves the best known sample complexity of AC by an order of $\mathcal{O}(\epsilon^{-1}\log(1/\epsilon))$, and the overall sample complexity for a mini-batch NAC to attain an $\epsilon$-accurate globally optimal point improves the existing sample complexity of NAC by an order of $\mathcal{O}(\epsilon^{-1}/\log(1/\epsilon))$.
no code implementations • 26 Feb 2020 • Yi Zhou, Zhe Wang, Kaiyi Ji, Yingbin Liang, Vahid Tarokh
Our APG-restart is designed to 1) allow for adopting flexible parameter restart schemes that cover many existing ones; 2) have a global sub-linear convergence rate in nonconvex and nonsmooth optimization; and 3) have guaranteed convergence to a critical point and have various types of asymptotic convergence rates depending on the parameterization of local geometry in nonconvex and nonsmooth optimization.
2 code implementations • 18 Feb 2020 • Kaiyi Ji, Junjie Yang, Yingbin Liang
As a popular meta-learning approach, the model-agnostic meta-learning (MAML) algorithm has been widely used due to its simplicity and effectiveness.
no code implementations • 17 Feb 2020 • Ziwei Guan, Kaiyi Ji, Donald J Bucci Jr, Timothy Y Hu, Joseph Palombo, Michael Liston, Yingbin Liang
This paper investigates the attack model where an adversary attacks with a certain probability at each round, and its attack value can be arbitrary and unbounded if it attacks.
no code implementations • 15 Feb 2020 • Huaqing Xiong, Tengyu Xu, Yingbin Liang, Wei zhang
Despite the wide applications of Adam in reinforcement learning (RL), the theoretical convergence of Adam-type RL algorithms has not been established.
no code implementations • ICLR 2020 • Tengyu Xu, Zhe Wang, Yi Zhou, Yingbin Liang
Furthermore, the variance error (for both i. i. d.\ and Markovian sampling) and the bias error (for Markovian sampling) of VRTD are significantly reduced by the batch size of variance reduction in comparison to those of vanilla TD.
no code implementations • NeurIPS 2019 • Zhe Wang, Kaiyi Ji, Yi Zhou, Yingbin Liang, Vahid Tarokh
SARAH and SPIDER are two recently developed stochastic variance-reduced algorithms, and SPIDER has been shown to achieve a near-optimal first-order oracle complexity in smooth nonconvex optimization.
no code implementations • 27 Oct 2019 • Kaiyi Ji, Zhe Wang, Yi Zhou, Yingbin Liang
Two types of zeroth-order stochastic algorithms have recently been designed for nonconvex optimization respectively based on the first-order techniques SVRG and SARAH/SPIDER.
no code implementations • ICML 2020 • Kaiyi Ji, Zhe Wang, Bowen Weng, Yi Zhou, Wei zhang, Yingbin Liang
In this paper, we propose a novel scheme, which eliminates backtracking line search but still exploits the information along optimization path by adapting the batch size via history stochastic gradients.
no code implementations • 29 Sep 2019 • Jayanth Regatti, Gaurav Tendolkar, Yi Zhou, Abhishek Gupta, Yingbin Liang
The performance of fully synchronized distributed systems has faced a bottleneck due to the big data trend, under which asynchronous distributed systems are becoming a major popularity due to their powerful scalability.
no code implementations • NeurIPS 2019 • Tengyu Xu, Shaofeng Zou, Yingbin Liang
Gradient-based temporal difference (GTD) algorithms are widely used in off-policy learning scenarios.
no code implementations • 25 Sep 2019 • Bowen Weng, Huaqing Xiong, Yingbin Liang, Wei zhang
Differently from the popular Deep Q-Network (DQN) learning, Alternating Q-learning (AltQ) does not fully fit a target Q-function at each iteration, and is generally known to be unstable and inefficient.
no code implementations • 7 Feb 2019 • Yi Zhou, Zhe Wang, Kaiyi Ji, Yingbin Liang, Vahid Tarokh
In this paper, we develop novel momentum schemes with flexible coefficient settings to accelerate SPIDER for nonconvex and nonsmooth composite optimization, and show that the resulting algorithms achieve the near-optimal gradient oracle complexity for achieving a generalized first-order stationary condition.
no code implementations • NeurIPS 2019 • Shaofeng Zou, Tengyu Xu, Yingbin Liang
For this fitted SARSA algorithm, we also provide its finite-sample analysis.
no code implementations • ICLR 2019 • Yi Zhou, Junjie Yang, Huishuai Zhang, Yingbin Liang, Vahid Tarokh
Stochastic gradient descent (SGD) has been found to be surprisingly effective in training a variety of deep neural networks.
no code implementations • 22 Nov 2018 • Qunwei Li, Bhavya Kailkhura, Rushil Anirudh, Yi Zhou, Yingbin Liang, Pramod Varshney
Despite the growing interest in generative adversarial networks (GANs), training GANs remains a challenging problem, both from a theoretical and a practical standpoint.
no code implementations • NeurIPS 2018 • Kaiyi Ji, Yingbin Liang
An important class of distance metrics proposed for training generative adversarial networks (GANs) is the integral probability metric (IPM), in which the neural net distance captures the practical GAN training via two neural networks.
1 code implementation • 25 Oct 2018 • Zhe Wang, Kaiyi Ji, Yi Zhou, Yingbin Liang, Vahid Tarokh
SARAH and SPIDER are two recently developed stochastic variance-reduced algorithms, and SPIDER has been shown to achieve a near-optimal first-order oracle complexity in smooth nonconvex optimization.
no code implementations • 9 Oct 2018 • Zhe Wang, Yi Zhou, Yingbin Liang, Guanghui Lan
However, such a successful acceleration technique has not yet been proposed for second-order algorithms in nonconvex optimization. In this paper, we apply the momentum scheme to cubic regularized (CR) Newton's method and explore the potential for acceleration.
no code implementations • 22 Aug 2018 • Zhe Wang, Yi Zhou, Yingbin Liang, Guanghui Lan
This note considers the inexact cubic-regularized Newton's method (CR), which has been shown in \cite{Cartis2011a} to achieve the same order-level convergence rate to a secondary stationary point as the exact CR \citep{Nesterov2006}.
no code implementations • NeurIPS 2018 • Yi Zhou, Zhe Wang, Yingbin Liang
Cubic-regularized Newton's method (CR) is a popular algorithm that guarantees to produce a second-order stationary solution for solving nonconvex optimization problems.
no code implementations • 31 Jul 2018 • Tiexing Wang, Qunwei Li, Donald J. Bucci, Yingbin Liang, Biao Chen, Pramod K. Varshney
In particular, the error exponent is characterized when either the Kolmogrov-Smirnov distance or the maximum mean discrepancy are used as the distance metric.
1 code implementation • ICLR 2019 • Tengyu Xu, Yi Zhou, Kaiyi Ji, Yingbin Liang
We study the implicit bias of gradient descent methods in solving a binary classification problem over a linearly separable dataset.
no code implementations • 20 Feb 2018 • Zhe Wang, Yi Zhou, Yingbin Liang, Guanghui Lan
Cubic regularization (CR) is an optimization method with emerging popularity due to its capability to escape saddle points and converge to second-order stationary solutions for nonconvex optimization.
no code implementations • 19 Feb 2018 • Yi Zhou, Yingbin Liang, Huishuai Zhang
With strongly convex regularizers, we further establish the generalization error bounds for nonconvex loss functions under proximal SGD with high-probability guarantee, i. e., exponential concentration in probability.
no code implementations • ICLR 2019 • Haoyu Fu, Yuejie Chi, Yingbin Liang
We prove that with Gaussian inputs, the empirical risk based on cross entropy exhibits strong convexity and smoothness {\em uniformly} in a local neighborhood of the ground truth, as soon as the sample complexity is sufficiently large.
no code implementations • ICLR 2018 • Yi Zhou, Yingbin Liang
In this paper, we provide a necessary and sufficient characterization of the analytical forms for the critical points (as well as global minimizers) of the square loss functions for linear neural networks.
no code implementations • 30 Oct 2017 • Yi Zhou, Yingbin Liang
We show that the analytical forms of the critical points characterize the values of the corresponding loss functions as well as the necessary and sufficient conditions to achieve global minimum.
no code implementations • 18 Oct 2017 • Yi Zhou, Yingbin Liang
The past decade has witnessed a successful application of deep learning to solving many challenging problems in machine learning and artificial intelligence.
no code implementations • 23 Sep 2017 • Yuanxin Li, Yuejie Chi, Huishuai Zhang, Yingbin Liang
Recent work has demonstrated the effectiveness of gradient descent for directly recovering the factors of low-rank matrices from random linear measurements in a globally convergent manner when initialized properly.
no code implementations • ICML 2017 • Qunwei Li, Yi Zhou, Yingbin Liang, Pramod K. Varshney
Then, by exploiting the Kurdyka-{\L}ojasiewicz (\KL) property for a broad class of functions, we establish the linear and sub-linear convergence rates of the function value sequence generated by APGnc.
no code implementations • NeurIPS 2016 • Huishuai Zhang, Yingbin Liang
In contrast to the smooth loss function used in WF, we adopt a nonsmooth but lower-order loss function, and design a gradient-like algorithm (referred to as reshaped-WF).
1 code implementation • 25 May 2016 • Huishuai Zhang, Yi Zhou, Yingbin Liang, Yuejie Chi
We further develop the incremental (stochastic) reshaped Wirtinger flow (IRWF) and show that IRWF converges linearly to the true signal.
no code implementations • 5 Apr 2016 • Shaofeng Zou, Yingbin Liang, H. Vincent Poor
Sufficient conditions on minimum and maximum sizes of candidate anomalous intervals are characterized in order to guarantee the proposed test to be consistent.
no code implementations • 11 Mar 2016 • Huishuai Zhang, Yuejie Chi, Yingbin Liang
This paper investigates the phase retrieval problem, which aims to recover a signal from the magnitudes of its linear measurements.
no code implementations • NeurIPS 2015 • Huishuai Zhang, Yi Zhou, Yingbin Liang
We investigate the robust PCA problem of decomposing an observed matrix into the sum of a low-rank and a sparse error matrices via convex programming Principal Component Pursuit (PCP).
no code implementations • 25 Apr 2014 • Shaofeng Zou, Yingbin Liang, H. Vincent Poor, Xinghua Shi
samples drawn from a distribution p, whereas each anomalous sequence contains m i. i. d.
no code implementations • 1 Apr 2014 • Shaofeng Zou, Yingbin Liang, H. Vincent Poor
If anomalous interval does not exist, then all nodes receive samples generated by p. It is assumed that the distributions p and q are arbitrary, and are unknown.
no code implementations • 30 Jul 2013 • Weiguang Wang, Yingbin Liang, Eric P. Xing
The goal is to recover the support union of all regression vectors using $l_1/l_2$-regularized Lasso.