no code implementations • 28 Nov 2023 • Dirk Tasche
Assumptions of invariance between the training joint distribution of features and labels and the test distribution can considerably facilitate this task.
no code implementations • 29 Mar 2023 • Dirk Tasche
We present new results on the transmission of SJS from sets of features to larger sets of features, a conditional correction formula for the class posterior probabilities under the target distribution, identifiability of SJS, and the relationship between SJS and covariate shift.
no code implementations • 29 Jul 2022 • Dirk Tasche
Factorizable joint shift (FJS) was recently proposed as a type of dataset shift for which the complete characteristics can be estimated from feature data observations on the test dataset by a method called Joint Importance Aligning.
no code implementations • 6 Jun 2022 • Dirk Tasche
We show that in the context of classification the property of source and target distributions to be related by covariate shift may be lost if the information content captured in the covariates is reduced, for instance by dropping components or mapping into a lower-dimensional or finite space.
no code implementations • 17 Jul 2021 • Dirk Tasche
For the binary prevalence quantification problem under prior probability shift, we determine the asymptotic variance of the maximum likelihood estimator.
no code implementations • 15 May 2021 • Dirk Tasche
When probabilistic classifiers are trained and calibrated, the so-called grouping loss component of the calibration loss can easily be overlooked.
no code implementations • 7 May 2020 • Dirk Tasche
We study how to perform tests on samples of pairs of observations and predictions in order to assess whether or not the predictions are prudent.
no code implementations • 10 Jun 2019 • Dirk Tasche
Less attention has been paid to the construction of confidence and prediction intervals for estimates of class prevalences.
no code implementations • 9 Apr 2018 • Dirk Tasche
For information retrieval and binary classification, we show that precision at the top (or precision at k) and recall at the top (or recall at k) are maximised by thresholding the posterior probability of the positive class.
1 code implementation • 19 Jan 2017 • Dirk Tasche
Lack of Fisher consistency could be used as a criterion to dismiss estimators that are unlikely to deliver precise estimates in test datasets under prior probability and more general dataset shift.
no code implementations • 28 Feb 2016 • Dirk Tasche
In contrast, quantification has been defined as the task of determining the prevalences of the different sorts of class labels in a target dataset.
no code implementations • 23 Jun 2014 • Dirk Tasche
A classical approach to this problem consists of fitting a mixture of the conditional score distributions observed last year to the current score distribution.
no code implementations • 2 Dec 2013 • Dirk Tasche
We quantify the bias of the total probability estimator of the unconditional class probabilities and show that the total odds estimator is unbiased.