1 code implementation • 17 Sep 2024 • Zilinghan Li, Shilan He, Ze Yang, Minseok Ryu, Kibaek Kim, Ravi Madduri
Federated learning (FL) is a distributed machine learning paradigm enabling collaborative model training while preserving data privacy.
1 code implementation • 17 Aug 2023 • Zilinghan Li, Shilan He, Pranshu Chaturvedi, Trung-Hieu Hoang, Minseok Ryu, E. A. Huerta, Volodymyr Kindratenko, Jordan Fuhrman, Maryellen Giger, Ryan Chard, Kibaek Kim, Ravi Madduri
Cross-silo privacy-preserving federated learning (PPFL) is a powerful tool to collaboratively train robust and generalized machine learning (ML) models without sharing sensitive (e. g., healthcare of financial) local data.
no code implementations • 28 Feb 2023 • Minseok Ryu, Kibaek Kim
This paper considers distributed optimization (DO) where multiple agents cooperate to minimize a global objective function, expressed as a sum of local objectives, subject to some constraints.
no code implementations • 18 Feb 2022 • Minseok Ryu, Kibaek Kim
Differential privacy (DP) techniques can be applied to the federated learning model to statistically guarantee data privacy against inference attacks to communication among the learning agents.
1 code implementation • 8 Feb 2022 • Minseok Ryu, Youngdae Kim, Kibaek Kim, Ravi K. Madduri
Federated learning (FL) enables training models at different sites and updating the weights from the training instead of transferring data to a central location and training as in classical machine learning.
no code implementations • 11 Jun 2021 • Minseok Ryu, Kibaek Kim
Differential privacy (DP) techniques can be applied to the federated learning model to protect data privacy against inference attacks to communication among the learning agents.