no code implementations • 8 Nov 2024 • Lucas Rosenblatt, Yuliia Lut, Eitan Turok, Marco Avella-Medina, Rachel Cummings
We consider DP variants of pre-processing methods that privately augment the original dataset to reduce the class imbalance; these include oversampling, SMOTE, and private synthetic data generation.
no code implementations • 20 Jul 2024 • Tingting Ou, Marco Avella Medina, Rachel Cummings
In this work we first show that the classical Thompson sampling algorithm for multi-arm bandits is differentially private as-is, without any modification.
no code implementations • 28 Jul 2023 • Rachel Cummings, Vitaly Feldman, Audra McMillan, Kunal Talwar
In this work we propose a simple model of heterogeneous user data that allows user data to differ in both distribution and quantity of data, and provide a method for estimating the population-level mean while preserving user-level differential privacy.
no code implementations • 24 Mar 2023 • Saeyoung Rho, Rachel Cummings, Vishal Misra
In this work, we provide the first algorithms for differentially private synthetic control with explicit error bounds.
1 code implementation • 2 Feb 2023 • Sloan Nietert, Rachel Cummings, Ziv Goldfeld
We prove new structural properties for POT and use them to show that MDE under a partial Wasserstein distance achieves the minimax-optimal robust estimation risk in many settings.
no code implementations • 10 Apr 2022 • Wanrong Zhang, Yajun Mei, Rachel Cummings
We also empirically validate our theoretical results on several synthetic databases, showing that our algorithms also perform well in practice.
1 code implementation • 2 Nov 2021 • Sloan Nietert, Rachel Cummings, Ziv Goldfeld
The Wasserstein distance, rooted in optimal transport (OT) theory, is a popular discrepancy measure between probability distributions with various applications to statistics and machine learning.
no code implementations • 25 Mar 2021 • Chris Waites, Rachel Cummings
Normalizing flow models have risen as a popular solution to the problem of density estimation, enabling high-quality synthetic data generation as well as exact probability density evaluation.
no code implementations • 24 Oct 2020 • Sebastian Perez-Salazar, Rachel Cummings
Our algorithms contains $k$ ordered experts that learn the best marginal item to select given the items chosen her predecessors, while maintaining privacy of the functions.
no code implementations • 8 Sep 2020 • Wanrong Zhang, Olga Ohrimenko, Rachel Cummings
We propose definitions to capture \emph{attribute privacy} in two relevant cases where global attributes may need to be protected: (1) properties of a specific dataset and (2) parameters of the underlying distribution from which dataset is sampled.
no code implementations • 20 Mar 2020 • Qiaomei Li, Rachel Cummings, Yonatan Mintz
In contrast to other heuristic methods, we use an integer optimization framework to combine local explainers into a near-global aggregate explainer.
1 code implementation • 27 Feb 2020 • Wanrong Zhang, Gautam Kamath, Rachel Cummings
In this work, we study False Discovery Rate (FDR) control in multiple hypothesis testing under the constraint of differential privacy for the sample.
9 code implementations • 10 Dec 2019 • Peter Kairouz, H. Brendan McMahan, Brendan Avent, Aurélien Bellet, Mehdi Bennis, Arjun Nitin Bhagoji, Kallista Bonawitz, Zachary Charles, Graham Cormode, Rachel Cummings, Rafael G. L. D'Oliveira, Hubert Eichner, Salim El Rouayheb, David Evans, Josh Gardner, Zachary Garrett, Adrià Gascón, Badih Ghazi, Phillip B. Gibbons, Marco Gruteser, Zaid Harchaoui, Chaoyang He, Lie He, Zhouyuan Huo, Ben Hutchinson, Justin Hsu, Martin Jaggi, Tara Javidi, Gauri Joshi, Mikhail Khodak, Jakub Konečný, Aleksandra Korolova, Farinaz Koushanfar, Sanmi Koyejo, Tancrède Lepoint, Yang Liu, Prateek Mittal, Mehryar Mohri, Richard Nock, Ayfer Özgür, Rasmus Pagh, Mariana Raykova, Hang Qi, Daniel Ramage, Ramesh Raskar, Dawn Song, Weikang Song, Sebastian U. Stich, Ziteng Sun, Ananda Theertha Suresh, Florian Tramèr, Praneeth Vepakomma, Jianyu Wang, Li Xiong, Zheng Xu, Qiang Yang, Felix X. Yu, Han Yu, Sen Zhao
FL embodies the principles of focused data collection and minimization, and can mitigate many of the systemic privacy risks and costs resulting from traditional, centralized machine learning and data science approaches.
1 code implementation • 6 Dec 2019 • Uthaipon Tantipongpipat, Chris Waites, Digvijay Boob, Amaresh Ankit Siva, Rachel Cummings
We introduce the DP-auto-GAN framework for synthetic data generation, which combines the low dimensional representation of autoencoders with the flexibility of Generative Adversarial Networks (GANs).
no code implementations • NeurIPS 2019 • Jacob D. Abernethy, Rachel Cummings, Bhuvesh Kumar, Sam Taggart, Jamie H. Morgenstern
We study the problem of learning Bayesian-optimal revenue-maximizing auctions.
no code implementations • ICML 2020 • Rachel Cummings, Sara Krehbiel, Yuliia Lut, Wanrong Zhang
Much of the prior work on change-point detection---including the only private algorithms for this problem---requires complete knowledge of the pre-change and post-change distributions.
1 code implementation • 25 Sep 2019 • Uthaipon Tantipongpipat, Chris Waites, Digvijay Boob, Amaresh Siva, Rachel Cummings
In this work we introduce the DP-auto-GAN framework for synthetic data generation, which combines the low dimensional representation of autoencoders with the flexibility of GANs.
no code implementations • NeurIPS 2018 • Rachel Cummings, Sara Krehbiel, Yajun Mei, Rui Tuo, Wanrong Zhang
The change-point detection problem seeks to identify distributional changes at an unknown change-point k* in a stream of data.
no code implementations • 6 Jul 2018 • Adrian Rivera Cardoso, Rachel Cummings
Our algorithm using bandit feedback is $\epsilon$-differentially private and achieves expected regret $\tilde{O}\left(\frac{n^{3/2}T^{3/4}}{\epsilon}\right)$.
no code implementations • 24 Feb 2016 • Rachel Cummings, Katrina Ligett, Kobbi Nissim, Aaron Roth, Zhiwei Steven Wu
We also show that perfect generalization is a strictly stronger guarantee than differential privacy, but that, nevertheless, many learning tasks can be carried out subject to the guarantees of perfect generalization.
no code implementations • 24 Feb 2016 • Rachel Cummings, David M. Pennock, Jennifer Wortman Vaughan
We consider the design of private prediction markets, financial markets designed to elicit predictions about uncertain events without revealing too much information about market participants' actions or beliefs.
no code implementations • 10 Jun 2015 • Rachel Cummings, Stratis Ioannidis, Katrina Ligett
We consider the problem of fitting a linear model to data held by individuals who are concerned about their privacy.
no code implementations • 27 Jul 2014 • Kareem Amin, Rachel Cummings, Lili Dworkin, Michael Kearns, Aaron Roth
We consider the problem of learning from revealed preferences in an online setting.