Improving Bayesian credibility intervals for classifier error rates using maximum entropy empirical priors
作者:
Highlights:
•
摘要
ObjectiveSuccessful use of classifiers that learn to make decisions from a set of patient examples require robust methods for performance estimation. Recently many promising approaches for determination of an upper bound for the error rate of a single classifier have been reported but the Bayesian credibility interval (CI) obtained from a conventional holdout test still delivers one of the tightest bounds. The conventional Bayesian CI becomes unacceptably large in real world applications where the test set sizes are less than a few hundred. The source of this problem is that fact that the CI is determined exclusively by the result on the test examples. In other words, there is no information at all provided by the uniform prior density distribution employed which reflects complete lack of prior knowledge about the unknown error rate. Therefore, the aim of the study reported here was to study a maximum entropy (ME) based approach to improved prior knowledge and Bayesian CIs, demonstrating its relevance for biomedical research and clinical practice.
论文关键词:Classifier design,Performance evaluation,Small sample learning,Decision support system,Diagnosis,Prognosis
论文评审过程:Received 18 June 2009, Revised 7 December 2009, Accepted 16 February 2010, Available online 29 March 2010.
论文官网地址:https://doi.org/10.1016/j.artmed.2010.02.004