no code implementations • 2 Oct 2023 • Filippo Galli, Catuscia Palamidessi, Tommaso Cucinotta
Training differentially private machine learning models requires constraining an individual's contribution to the optimization process.
no code implementations • 1 Sep 2023 • Filippo Galli, Kangsoo Jung, Sayan Biswas, Catuscia Palamidessi, Tommaso Cucinotta
FL was proposed as a stepping-stone towards privacy-preserving machine learning, but it has been shown vulnerable to issues such as leakage of private information, lack of personalization of the model, and the possibility of having a trained model that is fairer to some groups than to others.
no code implementations • 7 Jun 2022 • Filippo Galli, Sayan Biswas, Kangsoo Jung, Tommaso Cucinotta, Catuscia Palamidessi
To cope with the issue of protecting the privacy of the clients and allowing for personalized model training to enhance the fairness and utility of the system, we propose a method to provide group privacy guarantees exploiting some key properties of $d$-privacy which enables personalized models under the framework of FL.
1 code implementation • 3 Nov 2021 • Giacomo Lanciano, Filippo Galli, Tommaso Cucinotta, Davide Bacciu, Andrea Passarella
Cloud auto-scaling mechanisms are typically based on reactive automation rules that scale a cluster whenever some metric, e. g., the average CPU usage among instances, exceeds a predefined threshold.