no code implementations • 9 Nov 2021 • Aleksei Triastcyn, Matthias Reisser, Christos Louizos
Privacy and communication efficiency are important challenges in federated training of neural networks, and combining them is still an open problem.
no code implementations • 23 Apr 2021 • Mohammad Samragh, Hossein Hosseini, Aleksei Triastcyn, Kambiz Azarian, Joseph Soriaga, Farinaz Koushanfar
In our method, the edge device runs the model up to a split layer determined based on its computational capacity.
no code implementations • 16 Nov 2020 • Panayiotis Danassis, Aleksei Triastcyn, Boi Faltings
We introduce a practical and scalable algorithm (PALMA) for solving one of the fundamental problems of multi-agent systems -- finding matches and allocations -- in unboundedly large settings (e. g., resource allocation in urban environments, mobility-on-demand systems, etc.
no code implementations • 2 Mar 2020 • Aleksei Triastcyn, Boi Faltings
This paper considers the problem of enhancing user privacy in common machine learning development tasks, such as data annotation and inspection, by substituting the real data with samples form a generative adversarial network.
no code implementations • 22 Nov 2019 • Aleksei Triastcyn, Boi Faltings
We consider the problem of reinforcing federated learning with formal privacy guarantees.
no code implementations • 18 Oct 2019 • Aleksei Triastcyn, Boi Faltings
In this paper, we propose FedGP, a framework for privacy-preserving data release in the federated learning setting.
1 code implementation • ICML 2020 • Aleksei Triastcyn, Boi Faltings
Traditional differential privacy is independent of the data distribution.
no code implementations • 8 Mar 2018 • Aleksei Triastcyn, Boi Faltings
In this paper, we propose generating artificial data that retain statistical properties of real data as the means of providing privacy with respect to the original dataset.
BIG-bench Machine Learning Generative Adversarial Network +1
no code implementations • ICLR 2018 • Aleksei Triastcyn, Boi Faltings
In this paper, we present a technique for generating artificial datasets that retain statistical properties of the real data while providing differential privacy guarantees with respect to this data.
BIG-bench Machine Learning Generative Adversarial Network +1