no code implementations • 30 Jul 2020 • Ning Xu, Timothy C. G. Fisher, Jian Hong
We establish a general upper bound for $K$-fold cross-validation ($K$-CV) errors that can be adapted to many $K$-CV-based estimators and learning algorithms.
no code implementations • 30 Jul 2020 • Ning Xu, Timothy C. G. Fisher, Jian Hong
In this paper, we merge two well-known tools from machine learning and biostatistics---variable selection algorithms and probablistic graphs---to estimate house prices and the corresponding causal structure using 2010 data on Sydney.
no code implementations • 30 Jul 2020 • Ning Xu, Timothy C. G. Fisher, Jian Hong
In this paper we focus on the empirical variable-selection peformance of subsample-ordered least angle regression (Solar) -- a novel ultrahigh dimensional redesign of lasso -- on the empirical data with complicated dependence structures and, hence, severe multicollinearity and grouping effect issues.
no code implementations • 20 May 2017 • Ning Xu, Jian Hong, Timothy C. G. Fisher
The $\left( \beta, \varpi \right)$-stability mathematically connects the generalization ability and the stability of the cross-validated model via the Rademacher complexity.
no code implementations • 18 Oct 2016 • Ning Xu, Jian Hong, Timothy C. G. Fisher
We propose using generalization error minimization (GEM) as a framework for model selection.
no code implementations • 12 Sep 2016 • Ning Xu, Jian Hong, Timothy C. G. Fisher
We show that the error bounds may be used for tuning key estimation hyper-parameters, such as the number of folds $K$ in cross-validation.
no code implementations • 1 Jun 2016 • Ning Xu, Jian Hong, Timothy C. G. Fisher
In this paper, we study model selection from the perspective of generalization ability, under the framework of structural risk minimization (SRM) and Vapnik-Chervonenkis (VC) theory.