Search Results for author: Tomoya Murata

Found 10 papers, 0 papers with code

DIFF2: Differential Private Optimization via Gradient Differences for Nonconvex Distributed Learning

no code implementations8 Feb 2023 Tomoya Murata, Taiji Suzuki

In the previous work, the best known utility bound is $\widetilde O(\sqrt{d}/(n\varepsilon_\mathrm{DP}))$ in terms of the squared full gradient norm, which is achieved by Differential Private Gradient Descent (DP-GD) as an instance, where $n$ is the sample size, $d$ is the problem dimensionality and $\varepsilon_\mathrm{DP}$ is the differential privacy parameter.

Versatile Single-Loop Method for Gradient Estimator: First and Second Order Optimality, and its Application to Federated Learning

no code implementations1 Sep 2022 Kazusato Oko, Shunta Akiyama, Tomoya Murata, Taiji Suzuki

While variance reduction methods have shown great success in solving large scale optimization problems, many of them suffer from accumulated errors and, therefore, should periodically require the full gradient computation.

Federated Learning

Escaping Saddle Points with Bias-Variance Reduced Local Perturbed SGD for Communication Efficient Nonconvex Distributed Learning

no code implementations12 Feb 2022 Tomoya Murata, Taiji Suzuki

In recent centralized nonconvex distributed learning and federated learning, local methods are one of the promising approaches to reduce communication time.

Distributed Optimization Federated Learning

Bias-Variance Reduced Local SGD for Less Heterogeneous Federated Learning

no code implementations5 Feb 2021 Tomoya Murata, Taiji Suzuki

Recently, local SGD has got much attention and been extensively studied in the distributed learning community to overcome the communication bottleneck problem.

Distributed Optimization Federated Learning

Gradient Descent in RKHS with Importance Labeling

no code implementations19 Jun 2020 Tomoya Murata, Taiji Suzuki

In this paper, we study importance labeling problem, in which we are given many unlabeled data and select a limited number of data to be labeled from the unlabeled data, and then a learning algorithm is executed on the selected one.

Accelerated Sparsified SGD with Error Feedback

no code implementations29 May 2019 Tomoya Murata, Taiji Suzuki

Several work has shown that {\it{sparsified}} stochastic gradient descent method (SGD) with {\it{error feedback}} asymptotically achieves the same rate as (non-sparsified) parallel SGD.

Distributed Optimization

Sample Efficient Stochastic Gradient Iterative Hard Thresholding Method for Stochastic Sparse Linear Regression with Limited Attribute Observation

no code implementations NeurIPS 2018 Tomoya Murata, Taiji Suzuki

We develop new stochastic gradient methods for efficiently solving sparse linear regression in a partial attribute observation setting, where learners are only allowed to observe a fixed number of actively chosen attributes per example at training and prediction times.

Attribute

Doubly Accelerated Stochastic Variance Reduced Dual Averaging Method for Regularized Empirical Risk Minimization

no code implementations NeurIPS 2017 Tomoya Murata, Taiji Suzuki

In this paper, we develop a new accelerated stochastic gradient method for efficiently solving the convex regularized empirical risk minimization problem in mini-batch settings.

Stochastic dual averaging methods using variance reduction techniques for regularized empirical risk minimization problems

no code implementations8 Mar 2016 Tomoya Murata, Taiji Suzuki

We consider a composite convex minimization problem associated with regularized empirical risk minimization, which often arises in machine learning.

BIG-bench Machine Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.