Search Results for author: Ding-Xuan Zhou

Found 55 papers, 4 papers with code

Spherical Analysis of Learning Nonlinear Functionals

no code implementations1 Oct 2024 Zhenyu Yang, Shuo Huang, Han Feng, Ding-Xuan Zhou

It utilizes spherical harmonics to help us extract the latent finite-dimensional information of functions, which in turn facilitates in the next step of approximation analysis using fully connected neural networks.

Decoder

Convergence Analysis for Deep Sparse Coding via Convolutional Neural Networks

no code implementations10 Aug 2024 Jianfei Li, Han Feng, Ding-Xuan Zhou

In this work, we explore intersections between sparse coding and deep learning to enhance our understanding of feature extraction capabilities in advanced neural network architectures.

Deep Learning

Bridging Smoothness and Approximation: Theoretical Insights into Over-Smoothing in Graph Neural Networks

no code implementations1 Jul 2024 Guangrui Yang, Jianfei Li, Ming Li, Han Feng, Ding-Xuan Zhou

In our numerical experiments, we analyze several widely applied GCNs and observe the phenomenon of energy decay.

Generalization analysis with deep ReLU networks for metric and similarity learning

no code implementations10 May 2024 Junyu Zhou, Puyu Wang, Ding-Xuan Zhou

Specifically, by deriving the explicit form of the true metric for metric and similarity learning with the hinge loss, we construct a structured deep ReLU neural network as an approximation of the true metric, whose approximation ability relies on the network complexity.

On the rates of convergence for learning with convolutional neural networks

no code implementations25 Mar 2024 Yunfei Yang, Han Feng, Ding-Xuan Zhou

Our second result gives new analysis on the covering number of feed-forward neural networks with CNNs as special cases.

Binary Classification

Nonlinear functional regression by functional deep neural network with kernel embedding

no code implementations5 Jan 2024 Zhongjie Shi, Jun Fan, Linhao Song, Ding-Xuan Zhou, Johan A. K. Suykens

With the rapid development of deep learning in various fields of science and technology, such as speech recognition, image classification, and natural language processing, recently it is also widely applied in the functional data analysis (FDA) with some empirical success.

Dimensionality Reduction Image Classification +3

Lifting the Veil: Unlocking the Power of Depth in Q-learning

no code implementations27 Oct 2023 Shao-Bo Lin, Tao Li, Shaojie Tang, Yao Wang, Ding-Xuan Zhou

In this paper, we make fundamental contributions to the field of reinforcement learning by answering to the following three questions: Why does deep Q-learning perform so well?

Learning Theory Management +2

Adaptive Distributed Kernel Ridge Regression: A Feasible Distributed Learning Scheme for Data Silos

no code implementations8 Sep 2023 Di Wang, Xiaotong Liu, Shao-Bo Lin, Ding-Xuan Zhou

Data silos, mainly caused by privacy and interoperability, significantly constrain collaborations among different organizations with similar data for the same purpose.

Decision Making regression

Solving PDEs on Spheres with Physics-Informed Convolutional Neural Networks

no code implementations18 Aug 2023 Guanhang Lei, Zhen Lei, Lei Shi, Chenyu Zeng, Ding-Xuan Zhou

In this paper, we establish rigorous analysis of the physics-informed convolutional neural network (PICNN) for solving PDEs on the sphere.

Classification with Deep Neural Networks and Logistic Loss

no code implementations31 Jul 2023 Zihan Zhang, Lei Shi, Ding-Xuan Zhou

In this paper, we aim to fill this gap by establishing a novel and elegant oracle-type inequality, which enables us to deal with the boundedness restriction of the target function, and using it to derive sharp convergence rates for fully connected ReLU DNN classifiers trained with logistic loss.

Binary Classification Classification +1

Deep Convolutional Neural Networks with Zero-Padding: Feature Extraction and Learning

1 code implementation30 Jul 2023 Zhi Han, Baichen Liu, Shao-Bo Lin, Ding-Xuan Zhou

This paper studies the performance of deep convolutional neural networks (DCNNs) with zero-padding in feature extraction and learning.

Translation

Rates of Approximation by ReLU Shallow Neural Networks

no code implementations24 Jul 2023 Tong Mao, Ding-Xuan Zhou

We show that ReLU shallow neural networks with $m$ hidden neurons can uniformly approximate functions from the H\"older space $W_\infty^r([-1, 1]^d)$ with rates $O((\log m)^{\frac{1}{2} +d}m^{-\frac{r}{d}\frac{d+2}{d+4}})$ when $r<d/2 +2$.

Learning Theory of Distribution Regression with Neural Networks

no code implementations7 Jul 2023 Zhongjie Shi, Zhan Yu, Ding-Xuan Zhou

In contrast to the classical regression methods, the input variables of distribution regression are probability measures.

Learning Theory regression

Nonparametric regression using over-parameterized shallow ReLU neural networks

no code implementations14 Jun 2023 Yunfei Yang, Ding-Xuan Zhou

It is shown that over-parameterized neural networks can achieve minimax optimal rates of convergence (up to logarithmic factors) for learning functions from certain smooth function classes, if the weights are suitably constrained or regularized.

regression

Fine-grained analysis of non-parametric estimation for pairwise learning

no code implementations31 May 2023 Junyu Zhou, Shuo Huang, Han Feng, Puyu Wang, Ding-Xuan Zhou

In this paper, we are concerned with the generalization performance of non-parametric estimation for pairwise learning.

regression

Generalization Guarantees of Gradient Descent for Multi-Layer Neural Networks

no code implementations26 May 2023 Puyu Wang, Yunwen Lei, Di Wang, Yiming Ying, Ding-Xuan Zhou

This sheds light on sufficient or necessary conditions for under-parameterized and over-parameterized NNs trained by GD to attain the desired risk rate of $O(1/\sqrt{n})$.

Distributed Gradient Descent for Functional Learning

no code implementations12 May 2023 Zhan Yu, Jun Fan, Zhongjie Shi, Ding-Xuan Zhou

In the information era, to face the big data challenges {that} stem from functional data analysis very recently, we propose a novel distributed gradient descent functional learning (DGDFL) algorithm to tackle functional data across numerous local machines (processors) in the framework of reproducing kernel Hilbert space.

Approximation of Nonlinear Functionals Using Deep ReLU Networks

no code implementations10 Apr 2023 Linhao Song, Jun Fan, Di-Rong Chen, Ding-Xuan Zhou

In recent years, functional neural networks have been proposed and studied in order to approximate nonlinear continuous functionals defined on $L^p([-1, 1]^s)$ for integers $s\ge1$ and $1\le p<\infty$.

Optimal rates of approximation by shallow ReLU$^k$ neural networks and applications to nonparametric regression

no code implementations4 Apr 2023 Yunfei Yang, Ding-Xuan Zhou

It is also proven that over-parameterized (deep or shallow) neural networks can achieve nearly optimal rates for nonparametric regression.

regression

Sketching with Spherical Designs for Noisy Data Fitting on Spheres

no code implementations8 Mar 2023 Shao-Bo Lin, Di Wang, Ding-Xuan Zhou

These interesting findings show that the proposed sketching strategy is capable of fitting massive and noisy data on spheres.

Generalization Analysis for Contrastive Representation Learning

no code implementations24 Feb 2023 Yunwen Lei, Tianbao Yang, Yiming Ying, Ding-Xuan Zhou

For self-bounding Lipschitz loss functions, we further improve our results by developing optimistic bounds which imply fast rates in a low noise condition.

Contrastive Learning Generalization Bounds +1

SignReLU neural network and its approximation ability

1 code implementation19 Oct 2022 Jianfei Li, Han Feng, Ding-Xuan Zhou

Deep neural networks (DNNs) have garnered significant attention in various fields of science and technology in recent years.

Approximation analysis of CNNs from a feature extraction view

no code implementations14 Oct 2022 Jianfei Li, Han Feng, Ding-Xuan Zhou

In this paper we establish some analysis for linear feature extraction by a deep multi-channel convolutional neural networks (CNNs), which demonstrates the power of deep learning over traditional linear transformations, like Fourier, wavelets, redundant dictionary coding methods.

Stability and Generalization for Markov Chain Stochastic Gradient Methods

no code implementations16 Sep 2022 Puyu Wang, Yunwen Lei, Yiming Ying, Ding-Xuan Zhou

To the best of our knowledge, this is the first generalization analysis of SGMs when the gradients are sampled from a Markov process.

Generalization Bounds Learning Theory

Differentially Private Stochastic Gradient Descent with Low-Noise

no code implementations9 Sep 2022 Puyu Wang, Yunwen Lei, Yiming Ying, Ding-Xuan Zhou

In this paper, we focus on the privacy and utility (measured by excess risk bounds) performances of differentially private stochastic gradient descent (SGD) algorithms in the setting of stochastic convex optimization.

Privacy Preserving

Attention Enables Zero Approximation Error

no code implementations24 Feb 2022 Zhiying Fang, Yidong Ouyang, Ding-Xuan Zhou, Guang Cheng

In this work, we show that with suitable adaptations, the single-head self-attention transformer with a fixed number of transformer encoder blocks and free parameters is able to generate any desired polynomial of the input with no error.

Deep Learning Image Classification

Radial Basis Function Approximation with Distributively Stored Data on Spheres

no code implementations5 Dec 2021 Han Feng, Shao-Bo Lin, Ding-Xuan Zhou

This paper proposes a distributed weighted regularized least squares algorithm (DWRLS) based on spherical radial basis functions and spherical quadrature rules to tackle spherical data that are stored across numerous local servers and cannot be shared with each other.

Generalization Performance of Empirical Risk Minimization on Over-parameterized Deep ReLU Nets

no code implementations28 Nov 2021 Shao-Bo Lin, Yao Wang, Ding-Xuan Zhou

In this paper, we study the generalization performance of global minima for implementing empirical risk minimization (ERM) on over-parameterized deep ReLU nets.

Theory of Deep Convolutional Neural Networks III: Approximating Radial Functions

no code implementations2 Jul 2021 Tong Mao, Zhongjie Shi, Ding-Xuan Zhou

We consider a family of deep neural networks consisting of two groups of convolutional layers, a downsampling operator, and a fully connected layer.

regression

Universal Consistency of Deep Convolutional Neural Networks

no code implementations23 Jun 2021 Shao-Bo Lin, Kaidong Wang, Yao Wang, Ding-Xuan Zhou

Compared with avid research activities of deep convolutional neural networks (DCNNs) in practice, the study of theoretical behaviors of DCNNs lags heavily behind.

Robust Kernel-based Distribution Regression

no code implementations21 Apr 2021 Zhan Yu, Daniel W. C. Ho, Ding-Xuan Zhou

Regularization schemes for regression have been widely studied in learning theory and inverse problems.

Learning Theory regression

Moreau Envelope Augmented Lagrangian Method for Nonconvex Optimization with Linear Constraints

no code implementations21 Jan 2021 Jinshan Zeng, Wotao Yin, Ding-Xuan Zhou

We modify ALM to use a Moreau envelope of the augmented Lagrangian and establish its convergence under conditions that are weaker than those in the literature.

Optimization and Control

Theory of Deep Convolutional Neural Networks II: Spherical Analysis

no code implementations28 Jul 2020 Zhiying Fang, Han Feng, Shuo Huang, Ding-Xuan Zhou

Deep learning based on deep neural networks of various structures and architectures has been powerful in many practical applications, but it lacks enough theoretical verifications.

Depth Selection for Deep ReLU Nets in Feature Extraction and Generalization

no code implementations1 Apr 2020 Zhi Han, Siquan Yu, Shao-Bo Lin, Ding-Xuan Zhou

One of the most important challenge of deep learning is to figure out relations between a feature and the depth of deep neural networks (deep nets for short) to reflect the necessity of depth.

Feature Engineering Ingenuity +1

Distributed Kernel Ridge Regression with Communications

no code implementations27 Mar 2020 Shao-Bo Lin, Di Wang, Ding-Xuan Zhou

This paper focuses on generalization performance analysis for distributed algorithms in the framework of learning theory.

Learning Theory regression

Realization of spatial sparseness by deep ReLU nets with massive data

no code implementations16 Dec 2019 Charles K. Chui, Shao-Bo Lin, Bo Zhang, Ding-Xuan Zhou

The great success of deep learning poses urgent challenges for understanding its working mechanism and rationality.

Deep Learning Learning Theory

Towards Understanding the Spectral Bias of Deep Learning

no code implementations3 Dec 2019 Yuan Cao, Zhiying Fang, Yue Wu, Ding-Xuan Zhou, Quanquan Gu

An intriguing phenomenon observed during training neural networks is the spectral bias, which states that neural networks are biased towards learning less complex functions.

Deep Learning

Optimal Stochastic and Online Learning with Individual Iterates

no code implementations NeurIPS 2019 Yunwen Lei, Peng Yang, Ke Tang, Ding-Xuan Zhou

In this paper, we propose a theoretically sound strategy to select an individual iterate of the vanilla SCMD, which is able to achieve optimal rates for both convex and strongly convex problems in a non-smooth learning setting.

Sparse Learning

Fast Polynomial Kernel Classification for Massive Data

1 code implementation24 Nov 2019 Jinshan Zeng, Minrun Wu, Shao-Bo Lin, Ding-Xuan Zhou

In the era of big data, it is desired to develop efficient machine learning algorithms to tackle massive data challenges such as storage bottleneck, algorithmic scalability, and interpretability.

Classification General Classification

Distributed filtered hyperinterpolation for noisy data on the sphere

no code implementations6 Oct 2019 Shao-Bo Lin, Yu Guang Wang, Ding-Xuan Zhou

This paper develops distributed filtered hyperinterpolation for noisy data on the sphere, which assigns the data fitting task to multiple servers to find a good approximation of the mapping of input and output data.

Geophysics Model Selection

Deep Neural Networks for Rotation-Invariance Approximation and Learning

no code implementations3 Apr 2019 Charles K. Chui, Shao-Bo Lin, Ding-Xuan Zhou

Based on the tree architecture, the objective of this paper is to design deep neural networks with two or more hidden layers (called deep nets) for realization of radial functions so as to enable rotational invariance for near-optimal function approximation in an arbitrarily high dimensional Euclidian space.

On ADMM in Deep Learning: Convergence and Saturation-Avoidance

1 code implementation6 Feb 2019 Jinshan Zeng, Shao-Bo Lin, Yuan YAO, Ding-Xuan Zhou

In this paper, we develop an alternating direction method of multipliers (ADMM) for deep neural networks training with sigmoid-type activation functions (called \textit{sigmoid-ADMM pair}), mainly motivated by the gradient-free nature of ADMM in avoiding the saturation of sigmoid-type activations and the advantages of deep neural networks with sigmoid-type activations (called deep sigmoid nets) over their rectified linear unit (ReLU) counterparts (called deep ReLU nets) in terms of approximation.

Deep Learning

Universality of Deep Convolutional Neural Networks

no code implementations28 May 2018 Ding-Xuan Zhou

Deep learning has been widely applied and brought breakthroughs in speech recognition, computer vision, and many other domains.

Learning Theory Open-Ended Question Answering +2

Construction of neural networks for realization of localized deep learning

no code implementations9 Mar 2018 Charles K. Chui, Shao-Bo Lin, Ding-Xuan Zhou

The subject of deep learning has recently attracted users of machine learning from various disciplines, including: medical diagnosis and bioinformatics, financial market analysis and online advertisement, speech and handwriting recognition, computer vision and natural language processing, time series forecasting, and search engines.

Deep Learning Dimensionality Reduction +4

Convergence of Online Mirror Descent

no code implementations18 Feb 2018 Yunwen Lei, Ding-Xuan Zhou

The condition is $\lim_{t\to\infty}\eta_t=0, \sum_{t=1}^{\infty}\eta_t=\infty$ in the case of positive variances.

Total stability of kernel methods

no code implementations22 Sep 2017 Andreas Christmann, Dao-Hong Xiang, Ding-Xuan Zhou

However, the actually used kernel often depends on one or on a few hyperparameters or the kernel is even data dependent in a much more complicated manner.

Data-dependent Generalization Bounds for Multi-class Classification

no code implementations29 Jun 2017 Yunwen Lei, Urun Dogan, Ding-Xuan Zhou, Marius Kloft

In this paper, we study data-dependent generalization error bounds exhibiting a mild dependency on the number of classes, making them suitable for multi-class learning with a large number of label classes.

Classification General Classification +2

Distributed learning with regularized least squares

no code implementations11 Aug 2016 Shao-Bo Lin, Xin Guo, Ding-Xuan Zhou

We study distributed learning with the least squares regularization scheme in a reproducing kernel Hilbert space (RKHS).

On the Robustness of Regularized Pairwise Learning Methods Based on Kernels

no code implementations12 Oct 2015 Andreas Christmann, Ding-Xuan Zhou

Regularized empirical risk minimization including support vector machines plays an important role in machine learning theory.

BIG-bench Machine Learning Learning Theory

Iterative Regularization for Learning with Convex Loss Functions

no code implementations31 Mar 2015 Junhong Lin, Lorenzo Rosasco, Ding-Xuan Zhou

We consider the problem of supervised learning with convex loss functions and propose a new form of iterative regularization based on the subgradient method.

BIG-bench Machine Learning

Minimax Optimal Rates of Estimation in High Dimensional Additive Models: Universal Phase Transition

no code implementations10 Mar 2015 Ming Yuan, Ding-Xuan Zhou

We establish minimax optimal rates of convergence for estimation in a high dimensional additive model assuming that it is approximately sparse.

Additive models regression

Unregularized Online Learning Algorithms with General Loss Functions

no code implementations2 Mar 2015 Yiming Ying, Ding-Xuan Zhou

Firstly, we derive explicit convergence rates of the unregularized online learning algorithms for classification associated with a general gamma-activating loss (see Definition 1 in the paper).

Online Pairwise Learning Algorithms with Kernels

no code implementations25 Feb 2015 Yiming Ying, Ding-Xuan Zhou

In this paper, we study an online algorithm for pairwise learning with a least-square loss function in an unconstrained setting of a reproducing kernel Hilbert space (RKHS), which we refer to as the Online Pairwise lEaRning Algorithm (OPERA).

Metric Learning

Consistency Analysis of an Empirical Minimum Error Entropy Algorithm

no code implementations17 Dec 2014 Jun Fan, Ting Hu, Qiang Wu, Ding-Xuan Zhou

The error entropy consistency, which requires the error entropy of the learned function to approximate the minimum error entropy, is shown to be always true if the bandwidth parameter tends to 0 at an appropriate rate.

regression

Cannot find the paper you are looking for? You can Submit a new open access paper.