Search Results for author: Jiaoyang Huang

Found 16 papers, 4 papers with code

Convergence Analysis of Probability Flow ODE for Score-based Generative Models

1 code implementation15 Apr 2024 Daniel Zhengyu Huang, Jiaoyang Huang, Zhengjiang Lin

Score-based generative models have emerged as a powerful approach for sampling high-dimensional probability distributions.

Sampling via Gradient Flows in the Space of Probability Measures

no code implementations5 Oct 2023 Yifan Chen, Daniel Zhengyu Huang, Jiaoyang Huang, Sebastian Reich, Andrew M Stuart

Our third contribution is to study, and develop efficient algorithms based on Gaussian approximations of the gradient flows; this leads to an alternative to particle methods.

Variational Inference

High-dimensional SGD aligns with emerging outlier eigenspaces

no code implementations4 Oct 2023 Gerard Ben Arous, Reza Gheissari, Jiaoyang Huang, Aukosh Jagannath

We rigorously study the joint evolution of training dynamics via stochastic gradient descent (SGD) and the spectra of empirical Hessian and gradient matrices.

How Does Information Bottleneck Help Deep Learning?

1 code implementation30 May 2023 Kenji Kawaguchi, Zhun Deng, Xu Ji, Jiaoyang Huang

In this paper, we provide the first rigorous learning theory for justifying the benefit of information bottleneck in deep learning by mathematically relating information bottleneck to generalization errors.

Generalization Bounds Learning Theory

Robustness Implies Generalization via Data-Dependent Generalization Bounds

no code implementations27 Jun 2022 Kenji Kawaguchi, Zhun Deng, Kyle Luh, Jiaoyang Huang

This paper proves that robustness implies generalization via data-dependent generalization bounds.

Generalization Bounds

Understanding End-to-End Model-Based Reinforcement Learning Methods as Implicit Parameterization

no code implementations NeurIPS 2021 Clement Gehring, Kenji Kawaguchi, Jiaoyang Huang, Leslie Kaelbling

Estimating the per-state expected cumulative rewards is a critical aspect of reinforcement learning approaches, however the experience is obtained, but standard deep neural-network function-approximation methods are often inefficient in this setting.

Model-based Reinforcement Learning reinforcement-learning +1

Long Random Matrices and Tensor Unfolding

no code implementations19 Oct 2021 Gérard Ben Arous, Daniel Zhengyu Huang, Jiaoyang Huang

In this paper, we consider the singular values and singular vectors of low rank perturbations of large rectangular random matrices, in the regime the matrix is "long": we allow the number of rows (columns) to grow polynomially in the number of columns (rows).

Improve Unscented Kalman Inversion With Low-Rank Approximation and Reduced-Order Model

no code implementations21 Feb 2021 Daniel Z. Huang, Jiaoyang Huang

The unscented Kalman inversion (UKI) presented in [1] is a general derivative-free approach to solving the inverse problem.

Numerical Analysis Numerical Analysis Optimization and Control

Spectrum of Random $d$-regular Graphs Up to the Edge

no code implementations1 Feb 2021 Jiaoyang Huang, Horng-Tzer Yau

Consider the normalized adjacency matrices of random $d$-regular graphs on $N$ vertices with fixed degree $d\geq3$.

Probability Mathematical Physics Combinatorics Mathematical Physics 60B20, 05C80

Towards Understanding the Dynamics of the First-Order Adversaries

no code implementations ICML 2020 Zhun Deng, Hangfeng He, Jiaoyang Huang, Weijie J. Su

An acknowledged weakness of neural networks is their vulnerability to adversarial perturbations to the inputs.

Dynamics of Deep Neural Networks and Neural Tangent Hierarchy

no code implementations ICML 2020 Jiaoyang Huang, Horng-Tzer Yau

However, it was observed in [5] that there is a performance gap between the kernel regression using the limiting NTK and the deep neural networks.

regression

Gradient Descent Finds Global Minima for Generalizable Deep Neural Networks of Practical Sizes

no code implementations5 Aug 2019 Kenji Kawaguchi, Jiaoyang Huang

The theory developed in this paper only requires the practical degrees of over-parameterization unlike previous theories.

Every Local Minimum Value is the Global Minimum Value of Induced Model in Non-convex Machine Learning

no code implementations7 Apr 2019 Kenji Kawaguchi, Jiaoyang Huang, Leslie Pack Kaelbling

Furthermore, as special cases of our general results, this article improves or complements several state-of-the-art theoretical results on deep neural networks, deep residual networks, and overparameterized deep neural networks with a unified proof technique and novel geometric insights.

BIG-bench Machine Learning Representation Learning

Effect of Depth and Width on Local Minima in Deep Learning

no code implementations20 Nov 2018 Kenji Kawaguchi, Jiaoyang Huang, Leslie Pack Kaelbling

In this paper, we analyze the effects of depth and width on the quality of local minima, without strong over-parameterization and simplification assumptions in the literature.

Cannot find the paper you are looking for? You can Submit a new open access paper.