Two-sample testing
61 papers with code • 5 benchmarks • 1 datasets
In statistical hypothesis testing, a two-sample test is a test performed on the data of two random samples, each independently obtained from a different given population. The purpose of the test is to determine whether the difference between these two populations is statistically significant. The statistics used in two-sample tests can be used to solve many machine learning problems, such as domain adaptation, covariate shift and generative adversarial networks.
Most implemented papers
PacGAN: The power of two samples in generative adversarial networks
Generative adversarial networks (GANs) are innovative techniques for learning generative models of complex data distributions from samples.
Adversarial Sample Detection for Deep Neural Network through Model Mutation Testing
We thus first propose a measure of `sensitivity' and show empirically that normal samples and adversarial samples have distinguishable sensitivity.
hyppo: A Multivariate Hypothesis Testing Python Package
We introduce hyppo, a unified library for performing multivariate hypothesis testing, including independence, two-sample, and k-sample testing.
Generative Moment Matching Networks
We consider the problem of learning deep generative models from data.
Gaussian Differential Privacy
More precisely, the privacy guarantees of \emph{any} hypothesis testing based definition of privacy (including original DP) converges to GDP in the limit under composition.
Association of genomic subtypes of lower-grade gliomas with shape features automatically extracted by a deep learning algorithm
Based on automatic deep learning segmentations, we extracted three features which quantify two-dimensional and three-dimensional characteristics of the tumors.
Online Robust Principal Component Analysis with Change Point Detection
Robust PCA methods are typically batch algorithms which requires loading all observations into memory before processing.
Statistical Anomaly Detection via Composite Hypothesis Testing for Markov Models
Under Markovian assumptions, we leverage a Central Limit Theorem (CLT) for the empirical measure in the test statistic of the composite hypothesis Hoeffding test so as to establish weak convergence results for the test statistic, and, thereby, derive a new estimator for the threshold needed by the test.
Scalable and Efficient Hypothesis Testing with Random Forests
Throughout the last decade, random forests have established themselves as among the most accurate and popular supervised learning methods.
Comparing distributions: $\ell_1$ geometry improves kernel two-sample testing
Here, we show that $L^p$ distances (with $p\geq 1$) between these distribution representatives give metrics on the space of distributions that are well-behaved to detect differences between distributions as they metrize the weak convergence.