8 code implementations • 10 Dec 2019 • Peter Kairouz, H. Brendan McMahan, Brendan Avent, Aurélien Bellet, Mehdi Bennis, Arjun Nitin Bhagoji, Kallista Bonawitz, Zachary Charles, Graham Cormode, Rachel Cummings, Rafael G. L. D'Oliveira, Hubert Eichner, Salim El Rouayheb, David Evans, Josh Gardner, Zachary Garrett, Adrià Gascón, Badih Ghazi, Phillip B. Gibbons, Marco Gruteser, Zaid Harchaoui, Chaoyang He, Lie He, Zhouyuan Huo, Ben Hutchinson, Justin Hsu, Martin Jaggi, Tara Javidi, Gauri Joshi, Mikhail Khodak, Jakub Konečný, Aleksandra Korolova, Farinaz Koushanfar, Sanmi Koyejo, Tancrède Lepoint, Yang Liu, Prateek Mittal, Mehryar Mohri, Richard Nock, Ayfer Özgür, Rasmus Pagh, Mariana Raykova, Hang Qi, Daniel Ramage, Ramesh Raskar, Dawn Song, Weikang Song, Sebastian U. Stich, Ziteng Sun, Ananda Theertha Suresh, Florian Tramèr, Praneeth Vepakomma, Jianyu Wang, Li Xiong, Zheng Xu, Qiang Yang, Felix X. Yu, Han Yu, Sen Zhao
FL embodies the principles of focused data collection and minimization, and can mitigate many of the systemic privacy risks and costs resulting from traditional, centralized machine learning and data science approaches.
3 code implementations • 25 Sep 2021 • Ashkan Yousefpour, Igor Shilov, Alexandre Sablayrolles, Davide Testuggine, Karthik Prasad, Mani Malek, John Nguyen, Sayan Ghosh, Akash Bharadwaj, Jessica Zhao, Graham Cormode, Ilya Mironov
We introduce Opacus, a free, open-source PyTorch library for training deep learning models with differential privacy (hosted at opacus. ai).
1 code implementation • 6 Oct 2022 • Samuel Maddock, Graham Cormode, Tianhao Wang, Carsten Maple, Somesh Jha
There is great demand for scalable, secure, and efficient privacy-preserving machine learning models that can be trained over distributed data.
1 code implementation • ICLR 2022 • Lauren Watson, Chuan Guo, Graham Cormode, Alex Sablayrolles
The vulnerability of machine learning models to membership inference attacks has received much attention in recent years.
1 code implementation • 26 Jul 2022 • Karthik Prasad, Sayan Ghosh, Graham Cormode, Ilya Mironov, Ashkan Yousefpour, Pierre Stock
Cross-device Federated Learning is an increasingly popular machine learning setting to train a model by leveraging a large population of client devices with high privacy and security guarantees.
1 code implementation • 17 Jan 2011 • Smriti Bhagat, Graham Cormode, S. Muthukrishnan
When dealing with large graphs, such as those that arise in the context of online social networks, a subset of nodes may be labeled.
Social and Information Networks Physics and Society
1 code implementation • 30 Oct 2019 • Graham Cormode, Charlie Dickens
Scalable algorithms to solve optimization and regression tasks even approximately, are needed to work with large datasets.
no code implementations • 5 Oct 2017 • Yu Zhang, Srikanta Tirthapura, Graham Cormode
We study Bayesian networks, the workhorse of graphical models, and present a communication-efficient method for continuously learning and maintaining a Bayesian network model over data that is arriving as a distributed stream partitioned across multiple processors.
no code implementations • ICML 2018 • Charlie Dickens, Graham Cormode, David Woodruff
Work on approximate linear algebra has led to efficient distributed and streaming algorithms for problems such as approximate matrix multiplication, low rank approximation, and regression, primarily for the Euclidean norm $\ell_2$.
no code implementations • 5 Apr 2021 • Ziyue Huang, Yuan Qiu, Ke Yi, Graham Cormode
We study the fundamental problem of frequency estimation under both privacy and communication constraints, where the data is distributed among $k$ parties.
no code implementations • 12 Apr 2022 • Saeed Mahloujifar, Alexandre Sablayrolles, Graham Cormode, Somesh Jha
A common countermeasure against MI attacks is to utilize differential privacy (DP) during model training to mask the presence of individual examples.
no code implementations • Proceedings of the 2014 ACM SIGMOD International Conference on Management of Data 2014 • Jun Zhang, Graham Cormode, Cecilia M. Procopiuc, Divesh Srivastava, Xiaokui Xiao
Given a dataset D, PRIVBAYES first constructs a Bayesian network N , which (i) provides a succinct model of the correlations among the attributes in D and (ii) allows us to approximate the distribution of data in D using a set P of lowdimensional marginals of D. After that, PRIVBAYES injects noise into each marginal in P to ensure differential privacy, and then uses the noisy marginals and the Bayesian network to construct an approximation of the data distribution in D. Finally, PRIVBAYES samples tuples from the approximate distribution to construct a synthetic dataset, and then releases the synthetic data.
no code implementations • 22 Oct 2022 • Graham Cormode, Igor Markov
We address two major obstacles to practical use of supervised classifiers on distributed private data.
no code implementations • 11 Jan 2023 • Sayan Ghosh, Karthik Prasad, Xiaoliang Dai, Peizhao Zhang, Bichen Wu, Graham Cormode, Peter Vajda
The resulting family of pruned models can consistently obtain better performance than existing FBNetV3 models at the same level of computation, and thus provide state-of-the-art results when trading off between computational complexity and generalization performance on the ImageNet benchmark.
1 code implementation • 5 Oct 2023 • Samuel Maddock, Graham Cormode, Carsten Maple
In this work, we initiate the study of federated synthetic tabular data generation.