no code implementations • 1 Feb 2024 • Mark Bun, Gautam Kamath, Argyris Mouzakis, Vikrant Singhal
We give an example of a class of distributions that is learnable in total variation distance with a finite number of samples, but not learnable under $(\varepsilon, \delta)$-differential privacy.
no code implementations • 13 Apr 2023 • Vikrant Singhal
We present the first $\varepsilon$-differentially private, computationally efficient algorithm that estimates the means of product distributions over $\{0, 1\}^d$ accurately in total-variation distance, whilst attaining the optimal sample complexity to within polylogarithmic factors.
no code implementations • 30 Jan 2023 • Gautam Kamath, Argyris Mouzakis, Matthew Regehr, Vikrant Singhal, Thomas Steinke, Jonathan Ullman
The canonical algorithm for differentially private mean estimation is to first clip the samples to a bounded range and then add noise to their empirical mean.
1 code implementation • 16 Aug 2022 • Alex Bie, Gautam Kamath, Vikrant Singhal
We initiate the study of differentially private (DP) estimation with access to a small amount of public data.
no code implementations • 17 May 2022 • Gautam Kamath, Argyris Mouzakis, Vikrant Singhal
First, we provide tight lower bounds for private covariance estimation of Gaussian distributions.
no code implementations • 8 Nov 2021 • Gautam Kamath, Argyris Mouzakis, Vikrant Singhal, Thomas Steinke, Jonathan Ullman
We give the first polynomial-time, polynomial-sample, differentially private estimator for the mean and covariance of an arbitrary Gaussian distribution $\mathcal{N}(\mu,\Sigma)$ in $\mathbb{R}^d$.
no code implementations • NeurIPS 2021 • Vikrant Singhal, Thomas Steinke
Private data analysis suffers a costly curse of dimensionality.
no code implementations • 21 Feb 2020 • Gautam Kamath, Vikrant Singhal, Jonathan Ullman
We give new upper and lower bounds on the minimax sample complexity of differentially private mean estimation of distributions with bounded $k$-th moments.
no code implementations • NeurIPS 2019 • Gautam Kamath, Or Sheffet, Vikrant Singhal, Jonathan Ullman
Learning the parameters of Gaussian mixture models is a fundamental and widely studied problem with numerous applications.
no code implementations • 1 May 2018 • Gautam Kamath, Jerry Li, Vikrant Singhal, Jonathan Ullman
We present novel, computationally efficient, and differentially private algorithms for two fundamental high-dimensional learning problems: learning a multivariate Gaussian and learning a product distribution over the Boolean hypercube in total variation distance.