no code implementations • 9 Dec 2023 • Ding Huang, Jian Huang, Ting Li, Guohao Shen
We propose a conditional stochastic interpolation (CSI) approach to learning conditional distributions.
no code implementations • 27 Jun 2023 • Shanshan Song, Tong Wang, Guohao Shen, Yuanyuan Lin, Jian Huang
Our approach simultaneously estimates a regression function and a conditional generator using a generative learning framework, where a conditional generator is a function that can generate samples from a conditional distribution.
no code implementations • 19 May 2023 • Guohao Shen
In this paper, we investigate the complexity of feed-forward neural networks by examining the concept of functional equivalence, which suggests that different network parameterizations can lead to the same function.
no code implementations • 1 May 2023 • Guohao Shen, Yuling Jiao, Yuanyuan Lin, Jian Huang
We establish error bounds for simultaneously approximating $C^s$ smooth functions and their derivatives using RePU-activated deep neural networks.
no code implementations • 18 Oct 2022 • Wenlu Tang, Guohao Shen, Yuanyuan Lin, Jian Huang
We also derive non-asymptotic upper bounds for the difference of the lengths between the proposed non-crossing conformal prediction interval and the theoretically oracle prediction interval.
no code implementations • 21 Jul 2022 • Guohao Shen, Yuling Jiao, Yuanyuan Lin, Joel L. Horowitz, Jian Huang
We propose a penalized nonparametric approach to estimating the quantile regression process (QRP) in a nonseparable model using rectifier quadratic unit (ReQU) activated deep neural networks and introduce a novel penalty function to enforce non-crossing of quantile regression curves.
no code implementations • 1 May 2021 • Guohao Shen, Yuling Jiao, Yuanyuan Lin, Jian Huang
To establish these results, we derive an upper bound for the covering number for the class of general convolutional neural networks with a bias term in each convolutional layer, and derive new results on the approximation power of CNNs for any uniformly-continuous target functions.