Ayuda
Ir al contenido

Dialnet


Resumen de Tracking cross-validated estimates of prediction error as studies accumulate

Lo-Bin Chang, Donald Geman

  • In recent years, “reproducibility” has emerged as a key factor in evaluating x applications of statistics to the biomedical sciences, for example, learning predictors of disease phenotypes from high-throughput “omics” data. In particular, “validation” is undermined when error rates on newly acquired data are sharply higher than those originally reported. More precisely, when data are collected from m “studies” representing possibly different subphenotypes, more generally different mixtures of subphenotypes, the error rates in cross-study validation (CSV) are observed to be larger than those obtained in ordinary randomized cross-validation (RCV), although the “gap” seems to close as m increases. Whereas these findings are hardly surprising for a heterogenous underlying population, this discrepancy is then seen as a barrier to translational research. We provide a statistical formulation in the large-sample limit: studies themselves are modeled as components of a mixture and all error rates are optimal (Bayes) for a two-class problem. Our results cohere with the trends observed in practice and suggest what is likely to be observed with large samples and consistent density estimators, namely, that the CSV error rate exceeds the RCV error rates for any m, the latter (appropriately averaged) increases with m, and both converge to the optimal rate for the whole population


Fundación Dialnet

Dialnet Plus

  • Más información sobre Dialnet Plus