Search Results for author: Shunta Akiyama

Found 5 papers, 0 papers with code

Diffusion Models are Minimax Optimal Distribution Estimators

no code implementations3 Mar 2023 Kazusato Oko, Shunta Akiyama, Taiji Suzuki

While efficient distribution learning is no doubt behind the groundbreaking success of diffusion modeling, its theoretical guarantees are quite limited.

Versatile Single-Loop Method for Gradient Estimator: First and Second Order Optimality, and its Application to Federated Learning

no code implementations1 Sep 2022 Kazusato Oko, Shunta Akiyama, Tomoya Murata, Taiji Suzuki

While variance reduction methods have shown great success in solving large scale optimization problems, many of them suffer from accumulated errors and, therefore, should periodically require the full gradient computation.

Federated Learning

Excess Risk of Two-Layer ReLU Neural Networks in Teacher-Student Settings and its Superiority to Kernel Methods

no code implementations30 May 2022 Shunta Akiyama, Taiji Suzuki

While deep learning has outperformed other methods for various tasks, theoretical frameworks that explain its reason have not been fully established.

On Learnability via Gradient Method for Two-Layer ReLU Neural Networks in Teacher-Student Setting

no code implementations11 Jun 2021 Shunta Akiyama, Taiji Suzuki

Deep learning empirically achieves high performance in many applications, but its training dynamics has not been fully understood theoretically.

Benefit of deep learning with non-convex noisy gradient descent: Provable excess risk bound and superiority to kernel methods

no code implementations ICLR 2021 Taiji Suzuki, Shunta Akiyama

Establishing a theoretical analysis that explains why deep learning can outperform shallow learning such as kernel methods is one of the biggest issues in the deep learning literature.

Cannot find the paper you are looking for? You can Submit a new open access paper.