no code implementations • 1 Oct 2024 • Zhenyu Yang, Shuo Huang, Han Feng, Ding-Xuan Zhou
It utilizes spherical harmonics to help us extract the latent finite-dimensional information of functions, which in turn facilitates in the next step of approximation analysis using fully connected neural networks.
no code implementations • 10 Aug 2024 • Jianfei Li, Han Feng, Ding-Xuan Zhou
In this work, we explore intersections between sparse coding and deep learning to enhance our understanding of feature extraction capabilities in advanced neural network architectures.
no code implementations • 1 Jul 2024 • Guangrui Yang, Jianfei Li, Ming Li, Han Feng, Ding-Xuan Zhou
In our numerical experiments, we analyze several widely applied GCNs and observe the phenomenon of energy decay.
no code implementations • 10 May 2024 • Junyu Zhou, Puyu Wang, Ding-Xuan Zhou
Specifically, by deriving the explicit form of the true metric for metric and similarity learning with the hinge loss, we construct a structured deep ReLU neural network as an approximation of the true metric, whose approximation ability relies on the network complexity.
no code implementations • 25 Mar 2024 • Yunfei Yang, Han Feng, Ding-Xuan Zhou
Our second result gives new analysis on the covering number of feed-forward neural networks with CNNs as special cases.
no code implementations • 5 Jan 2024 • Zhongjie Shi, Jun Fan, Linhao Song, Ding-Xuan Zhou, Johan A. K. Suykens
With the rapid development of deep learning in various fields of science and technology, such as speech recognition, image classification, and natural language processing, recently it is also widely applied in the functional data analysis (FDA) with some empirical success.
no code implementations • 27 Oct 2023 • Shao-Bo Lin, Tao Li, Shaojie Tang, Yao Wang, Ding-Xuan Zhou
In this paper, we make fundamental contributions to the field of reinforcement learning by answering to the following three questions: Why does deep Q-learning perform so well?
no code implementations • 8 Sep 2023 • Di Wang, Xiaotong Liu, Shao-Bo Lin, Ding-Xuan Zhou
Data silos, mainly caused by privacy and interoperability, significantly constrain collaborations among different organizations with similar data for the same purpose.
no code implementations • 18 Aug 2023 • Guanhang Lei, Zhen Lei, Lei Shi, Chenyu Zeng, Ding-Xuan Zhou
In this paper, we establish rigorous analysis of the physics-informed convolutional neural network (PICNN) for solving PDEs on the sphere.
no code implementations • 31 Jul 2023 • Zihan Zhang, Lei Shi, Ding-Xuan Zhou
In this paper, we aim to fill this gap by establishing a novel and elegant oracle-type inequality, which enables us to deal with the boundedness restriction of the target function, and using it to derive sharp convergence rates for fully connected ReLU DNN classifiers trained with logistic loss.
1 code implementation • 30 Jul 2023 • Zhi Han, Baichen Liu, Shao-Bo Lin, Ding-Xuan Zhou
This paper studies the performance of deep convolutional neural networks (DCNNs) with zero-padding in feature extraction and learning.
no code implementations • 24 Jul 2023 • Tong Mao, Ding-Xuan Zhou
We show that ReLU shallow neural networks with $m$ hidden neurons can uniformly approximate functions from the H\"older space $W_\infty^r([-1, 1]^d)$ with rates $O((\log m)^{\frac{1}{2} +d}m^{-\frac{r}{d}\frac{d+2}{d+4}})$ when $r<d/2 +2$.
no code implementations • 7 Jul 2023 • Zhongjie Shi, Zhan Yu, Ding-Xuan Zhou
In contrast to the classical regression methods, the input variables of distribution regression are probability measures.
no code implementations • 14 Jun 2023 • Yunfei Yang, Ding-Xuan Zhou
It is shown that over-parameterized neural networks can achieve minimax optimal rates of convergence (up to logarithmic factors) for learning functions from certain smooth function classes, if the weights are suitably constrained or regularized.
no code implementations • 31 May 2023 • Junyu Zhou, Shuo Huang, Han Feng, Puyu Wang, Ding-Xuan Zhou
In this paper, we are concerned with the generalization performance of non-parametric estimation for pairwise learning.
no code implementations • 26 May 2023 • Puyu Wang, Yunwen Lei, Di Wang, Yiming Ying, Ding-Xuan Zhou
This sheds light on sufficient or necessary conditions for under-parameterized and over-parameterized NNs trained by GD to attain the desired risk rate of $O(1/\sqrt{n})$.
no code implementations • 12 May 2023 • Zhan Yu, Jun Fan, Zhongjie Shi, Ding-Xuan Zhou
In the information era, to face the big data challenges {that} stem from functional data analysis very recently, we propose a novel distributed gradient descent functional learning (DGDFL) algorithm to tackle functional data across numerous local machines (processors) in the framework of reproducing kernel Hilbert space.
no code implementations • 10 Apr 2023 • Linhao Song, Jun Fan, Di-Rong Chen, Ding-Xuan Zhou
In recent years, functional neural networks have been proposed and studied in order to approximate nonlinear continuous functionals defined on $L^p([-1, 1]^s)$ for integers $s\ge1$ and $1\le p<\infty$.
no code implementations • 4 Apr 2023 • Yunfei Yang, Ding-Xuan Zhou
It is also proven that over-parameterized (deep or shallow) neural networks can achieve nearly optimal rates for nonparametric regression.
no code implementations • 8 Mar 2023 • Shao-Bo Lin, Di Wang, Ding-Xuan Zhou
These interesting findings show that the proposed sketching strategy is capable of fitting massive and noisy data on spheres.
no code implementations • 24 Feb 2023 • Yunwen Lei, Tianbao Yang, Yiming Ying, Ding-Xuan Zhou
For self-bounding Lipschitz loss functions, we further improve our results by developing optimistic bounds which imply fast rates in a low noise condition.
1 code implementation • 19 Oct 2022 • Jianfei Li, Han Feng, Ding-Xuan Zhou
Deep neural networks (DNNs) have garnered significant attention in various fields of science and technology in recent years.
no code implementations • 14 Oct 2022 • Jianfei Li, Han Feng, Ding-Xuan Zhou
In this paper we establish some analysis for linear feature extraction by a deep multi-channel convolutional neural networks (CNNs), which demonstrates the power of deep learning over traditional linear transformations, like Fourier, wavelets, redundant dictionary coding methods.
no code implementations • 16 Sep 2022 • Puyu Wang, Yunwen Lei, Yiming Ying, Ding-Xuan Zhou
To the best of our knowledge, this is the first generalization analysis of SGMs when the gradients are sampled from a Markov process.
no code implementations • 9 Sep 2022 • Puyu Wang, Yunwen Lei, Yiming Ying, Ding-Xuan Zhou
In this paper, we focus on the privacy and utility (measured by excess risk bounds) performances of differentially private stochastic gradient descent (SGD) algorithms in the setting of stochastic convex optimization.
no code implementations • 24 Feb 2022 • Zhiying Fang, Yidong Ouyang, Ding-Xuan Zhou, Guang Cheng
In this work, we show that with suitable adaptations, the single-head self-attention transformer with a fixed number of transformer encoder blocks and free parameters is able to generate any desired polynomial of the input with no error.
no code implementations • 5 Dec 2021 • Han Feng, Shao-Bo Lin, Ding-Xuan Zhou
This paper proposes a distributed weighted regularized least squares algorithm (DWRLS) based on spherical radial basis functions and spherical quadrature rules to tackle spherical data that are stored across numerous local servers and cannot be shared with each other.
no code implementations • 28 Nov 2021 • Shao-Bo Lin, Yao Wang, Ding-Xuan Zhou
In this paper, we study the generalization performance of global minima for implementing empirical risk minimization (ERM) on over-parameterized deep ReLU nets.
no code implementations • 2 Jul 2021 • Tong Mao, Zhongjie Shi, Ding-Xuan Zhou
We consider a family of deep neural networks consisting of two groups of convolutional layers, a downsampling operator, and a fully connected layer.
no code implementations • 23 Jun 2021 • Shao-Bo Lin, Kaidong Wang, Yao Wang, Ding-Xuan Zhou
Compared with avid research activities of deep convolutional neural networks (DCNNs) in practice, the study of theoretical behaviors of DCNNs lags heavily behind.
no code implementations • 21 Apr 2021 • Zhan Yu, Daniel W. C. Ho, Ding-Xuan Zhou
Regularization schemes for regression have been widely studied in learning theory and inverse problems.
no code implementations • 21 Jan 2021 • Jinshan Zeng, Wotao Yin, Ding-Xuan Zhou
We modify ALM to use a Moreau envelope of the augmented Lagrangian and establish its convergence under conditions that are weaker than those in the literature.
Optimization and Control
no code implementations • 28 Jul 2020 • Zhiying Fang, Han Feng, Shuo Huang, Ding-Xuan Zhou
Deep learning based on deep neural networks of various structures and architectures has been powerful in many practical applications, but it lacks enough theoretical verifications.
no code implementations • 1 Apr 2020 • Zhi Han, Siquan Yu, Shao-Bo Lin, Ding-Xuan Zhou
One of the most important challenge of deep learning is to figure out relations between a feature and the depth of deep neural networks (deep nets for short) to reflect the necessity of depth.
no code implementations • 27 Mar 2020 • Shao-Bo Lin, Di Wang, Ding-Xuan Zhou
This paper focuses on generalization performance analysis for distributed algorithms in the framework of learning theory.
no code implementations • 16 Dec 2019 • Charles K. Chui, Shao-Bo Lin, Bo Zhang, Ding-Xuan Zhou
The great success of deep learning poses urgent challenges for understanding its working mechanism and rationality.
no code implementations • 3 Dec 2019 • Yuan Cao, Zhiying Fang, Yue Wu, Ding-Xuan Zhou, Quanquan Gu
An intriguing phenomenon observed during training neural networks is the spectral bias, which states that neural networks are biased towards learning less complex functions.
no code implementations • NeurIPS 2019 • Yunwen Lei, Peng Yang, Ke Tang, Ding-Xuan Zhou
In this paper, we propose a theoretically sound strategy to select an individual iterate of the vanilla SCMD, which is able to achieve optimal rates for both convex and strongly convex problems in a non-smooth learning setting.
1 code implementation • 24 Nov 2019 • Jinshan Zeng, Minrun Wu, Shao-Bo Lin, Ding-Xuan Zhou
In the era of big data, it is desired to develop efficient machine learning algorithms to tackle massive data challenges such as storage bottleneck, algorithmic scalability, and interpretability.
no code implementations • 6 Oct 2019 • Shao-Bo Lin, Yu Guang Wang, Ding-Xuan Zhou
This paper develops distributed filtered hyperinterpolation for noisy data on the sphere, which assigns the data fitting task to multiple servers to find a good approximation of the mapping of input and output data.
no code implementations • 3 Apr 2019 • Charles K. Chui, Shao-Bo Lin, Ding-Xuan Zhou
Based on the tree architecture, the objective of this paper is to design deep neural networks with two or more hidden layers (called deep nets) for realization of radial functions so as to enable rotational invariance for near-optimal function approximation in an arbitrarily high dimensional Euclidian space.
1 code implementation • 6 Feb 2019 • Jinshan Zeng, Shao-Bo Lin, Yuan YAO, Ding-Xuan Zhou
In this paper, we develop an alternating direction method of multipliers (ADMM) for deep neural networks training with sigmoid-type activation functions (called \textit{sigmoid-ADMM pair}), mainly motivated by the gradient-free nature of ADMM in avoiding the saturation of sigmoid-type activations and the advantages of deep neural networks with sigmoid-type activations (called deep sigmoid nets) over their rectified linear unit (ReLU) counterparts (called deep ReLU nets) in terms of approximation.
no code implementations • 28 May 2018 • Ding-Xuan Zhou
Deep learning has been widely applied and brought breakthroughs in speech recognition, computer vision, and many other domains.
no code implementations • 9 Mar 2018 • Charles K. Chui, Shao-Bo Lin, Ding-Xuan Zhou
The subject of deep learning has recently attracted users of machine learning from various disciplines, including: medical diagnosis and bioinformatics, financial market analysis and online advertisement, speech and handwriting recognition, computer vision and natural language processing, time series forecasting, and search engines.
no code implementations • 18 Feb 2018 • Yunwen Lei, Ding-Xuan Zhou
The condition is $\lim_{t\to\infty}\eta_t=0, \sum_{t=1}^{\infty}\eta_t=\infty$ in the case of positive variances.
no code implementations • 22 Sep 2017 • Andreas Christmann, Dao-Hong Xiang, Ding-Xuan Zhou
However, the actually used kernel often depends on one or on a few hyperparameters or the kernel is even data dependent in a much more complicated manner.
no code implementations • 29 Jun 2017 • Yunwen Lei, Urun Dogan, Ding-Xuan Zhou, Marius Kloft
In this paper, we study data-dependent generalization error bounds exhibiting a mild dependency on the number of classes, making them suitable for multi-class learning with a large number of label classes.
no code implementations • 11 Aug 2016 • Shao-Bo Lin, Xin Guo, Ding-Xuan Zhou
We study distributed learning with the least squares regularization scheme in a reproducing kernel Hilbert space (RKHS).
no code implementations • 12 Oct 2015 • Andreas Christmann, Ding-Xuan Zhou
Regularized empirical risk minimization including support vector machines plays an important role in machine learning theory.
no code implementations • 31 Mar 2015 • Junhong Lin, Lorenzo Rosasco, Ding-Xuan Zhou
We consider the problem of supervised learning with convex loss functions and propose a new form of iterative regularization based on the subgradient method.
no code implementations • 10 Mar 2015 • Ming Yuan, Ding-Xuan Zhou
We establish minimax optimal rates of convergence for estimation in a high dimensional additive model assuming that it is approximately sparse.
no code implementations • 2 Mar 2015 • Yiming Ying, Ding-Xuan Zhou
Firstly, we derive explicit convergence rates of the unregularized online learning algorithms for classification associated with a general gamma-activating loss (see Definition 1 in the paper).
no code implementations • 25 Feb 2015 • Yiming Ying, Ding-Xuan Zhou
In this paper, we study an online algorithm for pairwise learning with a least-square loss function in an unconstrained setting of a reproducing kernel Hilbert space (RKHS), which we refer to as the Online Pairwise lEaRning Algorithm (OPERA).
no code implementations • 17 Dec 2014 • Jun Fan, Ting Hu, Qiang Wu, Ding-Xuan Zhou
The error entropy consistency, which requires the error entropy of the learned function to approximate the minimum error entropy, is shown to be always true if the bandwidth parameter tends to 0 at an appropriate rate.
no code implementations • 14 May 2014 • Andreas Christmann, Ding-Xuan Zhou
Additive models play an important role in semiparametric statistics.