no code implementations • 18 Oct 2023 • Tianyu Zhang, Jing Lei
We propose a weighted rolling-validation procedure, an online variant of leave-one-out cross-validation, that costs minimal extra computation for many typical stochastic gradient descent estimators.
2 code implementations • 26 May 2023 • Hang Zhou, Jonas Mueller, Mayank Kumar, Jane-Ling Wang, Jing Lei
Noise plagues many numerical datasets, where the recorded values in the data may fail to match the true underlying values due to reasons including: erroneous sensors, data entry/processing mistakes, or imperfect human estimates.
no code implementations • 15 Feb 2021 • Purvasha Chakravarti, Mikael Kuusela, Jing Lei, Larry Wasserman
Here we instead investigate a model-independent method that does not make any assumptions about the signal and uses a semi-supervised classifier to detect the presence of the signal in the experimental data.
Applications High Energy Physics - Phenomenology Data Analysis, Statistics and Probability
2 code implementations • 19 Nov 2019 • Yixuan Qiu, Jing Lei, Kathryn Roeder
In this work we study sparse PCA based on the convex FPS formulation, and propose a new algorithm that is computationally efficient and applicable to large and high-dimensional data sets.
no code implementations • 27 Apr 2018 • Jing Lei
We provide upper bounds of the expected Wasserstein distance between a probability measure and its empirical version, generalizing recent results for finite dimensional Euclidean spaces and bounded functional spaces.
1 code implementation • 23 Mar 2017 • Jing Lei
Cross-validation is one of the most popular model selection methods in statistics and machine learning.
no code implementations • 2 Sep 2016 • Mauricio Sadinle, Jing Lei, Larry Wasserman
In most classification tasks there are observations that are ambiguous and therefore difficult to correctly label.
no code implementations • 8 May 2016 • Yu-Xiang Wang, Jing Lei, Stephen E. Fienberg
We define On-Average KL-Privacy and present its properties and connections to differential privacy, generalization and information-theoretic quantities including max-information and mutual information.
5 code implementations • 14 Apr 2016 • Jing Lei, Max G'Sell, Alessandro Rinaldo, Ryan J. Tibshirani, Larry Wasserman
In the spirit of reproducibility, all of our empirical results can also be easily (re)generated using this package.
no code implementations • 13 Feb 2016 • Yu-Xiang Wang, Jing Lei, Stephen E. Fienberg
In this paper, we propose a minimax framework for adaptive data analysis.
no code implementations • 23 Feb 2015 • Yu-Xiang Wang, Jing Lei, Stephen E. Fienberg
Lastly, we extend some of the results to the more practical $(\epsilon,\delta)$-differential privacy and establish the existence of a phase-transition on the class of problems that are approximately privately learnable with respect to how small $\delta$ needs to be.
no code implementations • 6 Nov 2014 • Jing Lei, Lingxue Zhu
We propose and analyze a generic method for community recovery in stochastic block models and degree corrected block models.
no code implementations • 6 Aug 2014 • Mattia Ciollaro, Christopher Genovese, Jing Lei, Larry Wasserman
We introduce the functional mean-shift algorithm, an iterative algorithm for estimating the local modes of a surrogate density from functional data.
no code implementations • 27 Jan 2014 • Jing Lei, Vincent Q. Vu
What can be said about the results of sparse PCA without assuming a sparse and unique truth?
no code implementations • 7 Dec 2013 • Jing Lei, Alessandro Rinaldo
We analyze the performance of spectral clustering for community extraction in stochastic block models.
no code implementations • NeurIPS 2013 • Vincent Q. Vu, Juhee Cho, Jing Lei, Karl Rohe
We propose a novel convex relaxation of sparse principal subspace estimation based on the convex hull of rank-$d$ projection matrices (the Fantope).
no code implementations • 26 Feb 2013 • Jing Lei, Alessandro Rinaldo, Larry Wasserman
This paper applies conformal prediction techniques to compute simultaneous prediction bands and clustering trees for functional data.
no code implementations • 2 Nov 2012 • Vincent Q. Vu, Jing Lei
We study sparse principal components analysis in high dimensions, where $p$ (the number of variables) can be much larger than $n$ (the number of observations), and analyze the problem of estimating the subspace spanned by the principal eigenvectors of the population covariance matrix.
no code implementations • NeurIPS 2011 • Jing Lei
This paper studies privacy preserving M-estimators using perturbed histograms.