Search Results for author: Christopher Musco

Found 29 papers, 6 papers with code

A Simple and Practical Method for Reducing the Disparate Impact of Differential Privacy

no code implementations18 Dec 2023 Lucas Rosenblatt, Julia Stoyanovich, Christopher Musco

Our theoretical results center on the private mean estimation problem, while our empirical results center on extensive experiments on private data synthesis to demonstrate the effectiveness of stratification on a variety of private mechanisms.

Improved Active Learning via Dependent Leverage Score Sampling

no code implementations8 Oct 2023 Atsushi Shimizu, Xiaoou Cheng, Christopher Musco, Jonathan Weare

We show how to obtain improved active learning methods in the agnostic (adversarial noise) setting by combining marginal leverage score sampling with non-independent sampling strategies that promote spatial coverage.

Active Learning Uncertainty Quantification

Moments, Random Walks, and Limits for Spectrum Approximation

no code implementations2 Jul 2023 Yujia Jin, Christopher Musco, Aaron Sidford, Apoorv Vikram Singh

We study lower bounds for the problem of approximating a one dimensional distribution given (noisy) measurements of its moments.

Dimensionality Reduction for General KDE Mode Finding

no code implementations30 May 2023 Xinyu Luo, Christopher Musco, Cas Widdershoven

There has been particular interest in efficient methods for solving the problem when $D$ is represented as a mixture model or kernel density estimate, although few algorithmic results with worst-case approximation and runtime guarantees are known.

Dimensionality Reduction

Active Learning for Single Neuron Models with Lipschitz Non-Linearities

no code implementations24 Oct 2022 Aarshvi Gajjar, Chinmay Hegde, Christopher Musco

Namely, we can collect samples via statistical \emph{leverage score sampling}, which has been shown to be near-optimal in other active learning scenarios.

Active Learning

Active Linear Regression for $\ell_p$ Norms and Beyond

no code implementations9 Nov 2021 Cameron Musco, Christopher Musco, David P. Woodruff, Taisuke Yasuda

By combining this with our techniques for $\ell_p$ regression, we obtain an active regression algorithm making $\tilde O(d^{1+\max\{1, p/2\}}/\mathrm{poly}(\epsilon))$ queries for such loss functions, including the Tukey and Huber losses, answering another question of [CD21].

Dimensionality Reduction Open-Ended Question Answering +1

Correlation Sketches for Approximate Join-Correlation Queries

no code implementations7 Apr 2021 Aécio Santos, Aline Bessa, Fernando Chirigati, Christopher Musco, Juliana Freire

The increasing availability of structured datasets, from Web tables and open-data portals to enterprise data, opens up opportunities~to enrich analytics and improve machine learning models through relational data augmentation.

Data Augmentation

The Statistical Cost of Robust Kernel Hyperparameter Turning

no code implementations NeurIPS 2020 Raphael Meyer, Christopher Musco

This paper studies the statistical complexity of kernel hyperparameter tuning in the setting of active regression under adversarial noise.

Hyperparameter Optimization regression

Hutch++: Optimal Stochastic Trace Estimation

1 code implementation19 Oct 2020 Raphael A. Meyer, Cameron Musco, Christopher Musco, David P. Woodruff

This improves on the ubiquitous Hutchinson's estimator, which requires $O(1/\epsilon^2)$ matrix-vector products.

Fast and Near-Optimal Diagonal Preconditioning

no code implementations4 Aug 2020 Arun Jambulapati, Jerry Li, Christopher Musco, Aaron Sidford, Kevin Tian

In this paper, we revisit the decades-old problem of how to best improve $\mathbf{A}$'s condition number by left or right diagonal rescaling.

Graph Learning for Inverse Landscape Genetics

no code implementations22 Jun 2020 Prathamesh Dharangutte, Christopher Musco

Our main contribution is an efficient algorithm for \emph{inverse landscape genetics}, which is the task of inferring this graph from measurements of genetic similarity at different locations (graph nodes).

Graph Learning

The Statistical Cost of Robust Kernel Hyperparameter Tuning

no code implementations14 Jun 2020 Raphael A. Meyer, Christopher Musco

This paper studies the statistical complexity of kernel hyperparameter tuning in the setting of active regression under adversarial noise.

Hyperparameter Optimization regression

Fourier Sparse Leverage Scores and Approximate Kernel Learning

no code implementations NeurIPS 2020 Tamás Erdélyi, Cameron Musco, Christopher Musco

Bounding Fourier sparse leverage scores under various measures is of pure mathematical interest in approximation theory, and our work extends existing results for the uniform measure [Erd17, CP19a].

Active Learning Open-Ended Question Answering

Projection-Cost-Preserving Sketches: Proof Strategies and Constructions

no code implementations17 Apr 2020 Cameron Musco, Christopher Musco

In this note we illustrate how common matrix approximation methods, such as random projection and random sampling, yield projection-cost-preserving sketches, as introduced in [FSS13, CEM+15].

Sample Efficient Toeplitz Covariance Estimation

no code implementations14 May 2019 Yonina C. Eldar, Jerry Li, Cameron Musco, Christopher Musco

In addition to results that hold for any Toeplitz $T$, we further study the important setting when $T$ is close to low-rank, which is often the case in practice.

Simple Heuristics Yield Provable Algorithms for Masked Low-Rank Approximation

no code implementations22 Apr 2019 Cameron Musco, Christopher Musco, David P. Woodruff

In particular, for rank $k' > k$ depending on the $public\ coin\ partition\ number$ of $W$, the heuristic outputs rank-$k'$ $L$ with cost$(L) \leq OPT + \epsilon \|A\|_F^2$.

Low-Rank Matrix Completion Tensor Decomposition

A Universal Sampling Method for Reconstructing Signals with Simple Fourier Transforms

no code implementations20 Dec 2018 Haim Avron, Michael Kapralov, Cameron Musco, Christopher Musco, Ameya Velingker, Amir Zandieh

We formalize this intuition by showing that, roughly, a continuous signal from a given class can be approximately reconstructed using a number of samples proportional to the *statistical dimension* of the allowed power spectrum of that class.

Inferring Networks From Random Walk-Based Node Similarities

1 code implementation NeurIPS 2018 Jeremy Hoskins, Cameron Musco, Christopher Musco, Babis Tsourakakis

In this work we consider a privacy threat to a social network in which an attacker has access to a subset of random walk-based node similarities, such as effective resistances (i. e., commute times) or personalized PageRank scores.

Anomaly Detection Clustering +3

Random Fourier Features for Kernel Ridge Regression: Approximation Bounds and Statistical Guarantees

no code implementations ICML 2017 Haim Avron, Michael Kapralov, Cameron Musco, Christopher Musco, Ameya Velingker, Amir Zandieh

Qualitatively, our results are twofold: on the one hand, we show that random Fourier feature approximation can provably speed up kernel ridge regression under reasonable assumptions.

regression

Learning Networks from Random Walk-Based Node Similarities

1 code implementation23 Jan 2018 Jeremy G. Hoskins, Cameron Musco, Christopher Musco, Charalampos E. Tsourakakis

In this work we consider a privacy threat to a social network in which an attacker has access to a subset of random walk-based node similarities, such as effective resistances (i. e., commute times) or personalized PageRank scores.

Anomaly Detection Clustering +3

Minimizing Polarization and Disagreement in Social Networks

3 code implementations28 Dec 2017 Cameron Musco, Christopher Musco, Charalampos E. Tsourakakis

We perform an empirical study of our proposed methods on synthetic and real-world data that verify their value as mining tools to better understand the trade-off between of disagreement and polarization.

Recommendation Systems

Recursive Sampling for the Nystrom Method

no code implementations NeurIPS 2017 Cameron Musco, Christopher Musco

We give the first algorithm for kernel Nystrom approximation that runs in linear time in the number of training points and is provably accurate for all kernel matrices, without dependence on regularity or incoherence conditions.

Stability of the Lanczos Method for Matrix Function Approximation

1 code implementation25 Aug 2017 Cameron Musco, Christopher Musco, Aaron Sidford

In exact arithmetic, the method's error after $k$ iterations is bounded by the error of the best degree-$k$ polynomial uniformly approximating $f(x)$ on the range $[\lambda_{min}(A), \lambda_{max}(A)]$.

Data Structures and Algorithms Numerical Analysis

Recursive Sampling for the Nyström Method

2 code implementations24 May 2016 Cameron Musco, Christopher Musco

We give the first algorithm for kernel Nystr\"om approximation that runs in *linear time in the number of training points* and is provably accurate for all kernel matrices, without dependence on regularity or incoherence conditions.

Principal Component Projection Without Principal Component Analysis

no code implementations22 Feb 2016 Roy Frostig, Cameron Musco, Christopher Musco, Aaron Sidford

To achieve our results, we first observe that ridge regression can be used to obtain a "smooth projection" onto the top principal components.

regression

Input Sparsity Time Low-Rank Approximation via Ridge Leverage Score Sampling

no code implementations23 Nov 2015 Michael B. Cohen, Cameron Musco, Christopher Musco

Our method is based on a recursive sampling scheme for computing a representative subset of $A$'s columns, which is then used to find a low-rank approximation.

Randomized Block Krylov Methods for Stronger and Faster Approximate Singular Value Decomposition

no code implementations NeurIPS 2015 Cameron Musco, Christopher Musco

We give the first provable runtime improvement on Simultaneous Iteration: a simple randomized block Krylov method, closely related to the classic Block Lanczos algorithm, gives the same guarantees in just $\tilde{O}(1/\sqrt{\epsilon})$ iterations and performs substantially better experimentally.

Dimensionality Reduction for k-Means Clustering and Low Rank Approximation

no code implementations24 Oct 2014 Michael B. Cohen, Sam Elder, Cameron Musco, Christopher Musco, Madalina Persu

We show how to approximate a data matrix $\mathbf{A}$ with a much smaller sketch $\mathbf{\tilde A}$ that can be used to solve a general class of constrained k-rank approximation problems to within $(1+\epsilon)$ error.

Clustering Dimensionality Reduction

Uniform Sampling for Matrix Approximation

no code implementations21 Aug 2014 Michael B. Cohen, Yin Tat Lee, Cameron Musco, Christopher Musco, Richard Peng, Aaron Sidford

In addition to an improved understanding of uniform sampling, our main proof introduces a structural result of independent interest: we show that every matrix can be made to have low coherence by reweighting a small subset of its rows.

regression

Cannot find the paper you are looking for? You can Submit a new open access paper.