Search Results for author: Yuanxiong Guo

Found 8 papers, 0 papers with code

Communication and Energy Efficient Wireless Federated Learning with Intrinsic Privacy

no code implementations15 Apr 2023 Zhenxiao Zhang, Yuanxiong Guo, Yuguang Fang, Yanmin Gong

In this paper, we propose a novel wireless FL scheme called private federated edge learning with sparsification (PFELS) to provide client-level DP guarantee with intrinsic channel noise while reducing communication and energy overhead and improving model accuracy.

Federated Learning

Scalable and Low-Latency Federated Learning with Cooperative Mobile Edge Networking

no code implementations25 May 2022 Zhenxiao Zhang, Zhidong Gao, Yuanxiong Guo, Yanmin Gong

On the other hand, the edge-based FL framework that relies on an edge server co-located with mobile base station for model aggregation has low communication latency but suffers from degraded model accuracy due to the limited coverage of edge server.

Federated Learning

Federated Learning with Sparsified Model Perturbation: Improving Accuracy under Client-Level Differential Privacy

no code implementations15 Feb 2022 Rui Hu, Yanmin Gong, Yuanxiong Guo

Federated learning (FL) that enables edge devices to collaboratively learn a shared model while keeping their training data locally has received great attention recently and can protect privacy in comparison with the traditional centralized learning paradigm.

Federated Learning

Hybrid Local SGD for Federated Learning with Heterogeneous Communications

no code implementations ICLR 2022 Yuanxiong Guo, Ying Sun, Rui Hu, Yanmin Gong

Communication is a key bottleneck in federated learning where a large number of edge devices collaboratively learn a model under the orchestration of a central server without sharing their own training data.

Federated Learning

Federated Learning with Sparsification-Amplified Privacy and Adaptive Optimization

no code implementations1 Aug 2020 Rui Hu, Yanmin Gong, Yuanxiong Guo

Since sparsification would increase the number of communication rounds required to achieve a certain target accuracy, which is unfavorable for DP guarantee, we further introduce acceleration techniques to help reduce the privacy cost.

Federated Learning

Concentrated Differentially Private and Utility Preserving Federated Learning

no code implementations30 Mar 2020 Rui Hu, Yuanxiong Guo, Yanmin Gong

Federated learning is a machine learning setting where a set of edge devices collaboratively train a model under the orchestration of a central server without sharing their local data.

Federated Learning Privacy Preserving

Differentially Private Federated Learning for Resource-Constrained Internet of Things

no code implementations28 Mar 2020 Rui Hu, Yuanxiong Guo, E. Paul. Ratazzi, Yanmin Gong

With the proliferation of smart devices having built-in sensors, Internet connectivity, and programmable computation capability in the era of Internet of things (IoT), tremendous data is being generated at the network edge.

Federated Learning

DP-ADMM: ADMM-based Distributed Learning with Differential Privacy

no code implementations30 Aug 2018 Zonghao Huang, Rui Hu, Yuanxiong Guo, Eric Chan-Tin, Yanmin Gong

The goal of this paper is to provide differential privacy for ADMM-based distributed machine learning.

BIG-bench Machine Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.