1 code implementation • 11 Nov 2024 • Alex Havrilla, Wenjing Liao
Our theory predicts a power law between the generalization error and both the training data size and the network size for transformers, where the power depends on the intrinsic dimension $d$ of the training data.
no code implementations • 1 Oct 2024 • Hao liu, Zecheng Zhang, Wenjing Liao, Hayden Schaeffer
We establish a theoretical framework to quantify the neural scaling laws by analyzing its approximation and generalization errors.
no code implementations • 8 Jun 2024 • Hao liu, Jiahui Cheng, Wenjing Liao
This class encompasses a range of function types, such as functions with uniform regularity and discontinuous functions.
no code implementations • 19 Jan 2024 • Hao liu, Biraj Dahal, Rongjie Lai, Wenjing Liao
The problem of operator learning, in this context, seeks to extract these physical processes from empirical data, which is challenging due to the infinite or high dimensionality of data.
1 code implementation • 30 Nov 2023 • Alex Havrilla, Kevin Rojas, Wenjing Liao, Molei Tao
Diffusion generative models have achieved remarkable success in generating images with a fixed resolution.
no code implementations • 26 Jun 2023 • Zixuan Zhang, Minshuo Chen, Mengdi Wang, Wenjing Liao, Tuo Zhao
Existing theories on deep nonparametric regression have shown that when the input data lie on a low-dimensional manifold, deep neural networks can adapt to the intrinsic data structures.
no code implementations • 17 Mar 2023 • Hao liu, Alex Havrilla, Rongjie Lai, Wenjing Liao
Our paper establishes statistical guarantees on the generalization error of chart autoencoders, and we demonstrate their denoising capabilities by considering $n$ noisy training samples, along with their noise-free counterparts, on a $d$-dimensional manifold.
no code implementations • 25 Feb 2023 • Biraj Dahal, Alex Havrilla, Minshuo Chen, Tuo Zhao, Wenjing Liao
Many existing experiments have demonstrated that generative networks can generate high-dimensional complex data from a low-dimensional easy-to-sample distribution.
no code implementations • 1 Dec 2022 • Jiahui Cheng, Minshuo Chen, Hao liu, Tuo Zhao, Wenjing Liao
Label Shift has been widely believed to be harmful to the generalization performance of machine learning models.
1 code implementation • 6 Nov 2022 • Mengyi Tang, Wenjing Liao, Rachel Kuske, Sung Ha Kang
We propose a general and robust framework to recover differential equations using a weak formulation, for both ordinary and partial differential equations (ODEs and PDEs).
no code implementations • 9 Jun 2022 • Hao liu, Minshuo Chen, Siawpeng Er, Wenjing Liao, Tong Zhang, Tuo Zhao
Overparameterized neural networks enjoy great representation power on complex data, and more importantly yield sufficiently smooth output, which is crucial to their generalization and robustness.
no code implementations • 4 May 2022 • Jie Wang, Minshuo Chen, Tuo Zhao, Wenjing Liao, Yao Xie
Based on the approximation theory of neural networks, we show that the neural network IPM test has the type-II risk in the order of $n^{-(s+\beta)/d}$, which is in the same order of the type-II risk as the H\"older IPM test.
no code implementations • 1 Jan 2022 • Hao liu, Haizhao Yang, Minshuo Chen, Tuo Zhao, Wenjing Liao
Learning operators between infinitely dimensional spaces is an important learning task arising in wide applications in machine learning, imaging science, mathematical modeling and simulations, etc.
no code implementations • 7 Sep 2021 • Hao liu, Minshuo Chen, Tuo Zhao, Wenjing Liao
Most of existing statistical theories on deep neural networks have sample complexities cursed by the data dimension and therefore cannot well explain the empirical success of deep learning on high-dimensional data.
no code implementations • 13 Jan 2021 • Wenjing Liao, Mauro Maggioni, Stefano Vigogna
We consider the regression problem of estimating functions on $\mathbb{R}^D$ but supported on a $d$-dimensional manifold $ \mathcal{M} \subset \mathbb{R}^D $ with $ d \ll D $.
no code implementations • 3 Nov 2020 • Minshuo Chen, Hao liu, Wenjing Liao, Tuo Zhao
Our theory shows that deep neural networks are adaptive to the low-dimensional geometric structures of the covariates, and partially explains the success of deep learning for causal inference.
no code implementations • 10 Feb 2020 • Minshuo Chen, Wenjing Liao, Hongyuan Zha, Tuo Zhao
Generative Adversarial Networks (GANs) have achieved a great success in unsupervised learning.
no code implementations • 22 Jan 2020 • Hao Liu, Wenjing Liao
The estimation error of this variance quantity is also given in this paper.
no code implementations • NeurIPS 2019 • Minshuo Chen, Haoming Jiang, Wenjing Liao, Tuo Zhao
The network size scales exponentially in the approximation error, with an exponent depending on the intrinsic dimension of the data and the smoothness of the function.
no code implementations • NeurIPS 2019 • Minshuo Chen, Haoming Jiang, Wenjing Liao, Tuo Zhao
It therefore demonstrates the adaptivity of deep ReLU networks to low-dimensional geometric structures of data, and partially explains the power of deep ReLU networks in tackling high-dimensional data with low-dimensional geometric structures.
no code implementations • 6 Apr 2019 • Sung Ha Kang, Wenjing Liao, Yingjie Liu
The new algorithm, called Identifying Differential Equations with Numerical Time evolution (IDENT), is explored for data with non-periodic boundary conditions, noisy data and PDEs with varying coefficients.
Numerical Analysis
no code implementations • 3 Nov 2016 • Wenjing Liao, Mauro Maggioni
We consider the problem of efficiently approximating and encoding high-dimensional data sampled from a probability distribution $\rho$ in $\mathbb{R}^D$, that is nearly supported on a $d$-dimensional set $\mathcal{M}$ - for example supported on a $d$-dimensional Riemannian manifold.