no code implementations • 6 Feb 2024 • Xiaoxin Su, Yipeng Zhou, Laizhong Cui, John C. S. Lui, Jiangchuan Liu
In Federated Learning (FL) paradigm, a parameter server (PS) concurrently communicates with distributed participating clients for model collection, update aggregation, and model distribution over multiple rounds, without touching private data owned by individual clients.
no code implementations • 13 Dec 2021 • Laizhong Cui, Xiaoxin Su, Yipeng Zhou, Jiangchuan Liu
Federated Learning (FL) incurs high communication overhead, which can be greatly alleviated by compression for model updates.
no code implementations • 27 Dec 2020 • Gaoyang Liu, Xiaoqiang Ma, Yang Yang, Chen Wang, Jiangchuan Liu
In this paper, we take the first step to fill this gap by presenting FedEraser, the first federated unlearning methodology that can eliminate the influence of a federated client's data on the global FL model while significantly reducing the time used for constructing the unlearned FL model. The basic idea of FedEraser is to trade the central server's storage for unlearned model's construction time, where FedEraser reconstructs the unlearned model by leveraging the historical parameter updates of federated clients that have been retained at the central server during the training process of FL.
1 code implementation • 7 Jul 2020 • Yutao Huang, Lingyang Chu, Zirui Zhou, Lanjun Wang, Jiangchuan Liu, Jian Pei, Yong Zhang
Non-IID data present a tough challenge for federated learning.
no code implementations • 22 Mar 2014 • Jie Xu, Mihaela van der Schaar, Jiangchuan Liu, Haitao Li
This paper presents a systematic online prediction method (Social-Forecast) that is capable to accurately forecast the popularity of videos promoted by social media.