A one-dimensional analysis for the probability of error of linear classifiers for normally distributed classes

作者:

Highlights:

摘要

Computing the probability of error is an important problem in evaluating classifiers. When dealing with normally distributed classes, this problem becomes intricate due to the fact that there is no closed-form expression for integrating the probability density function. In this paper, we derive lower and upper bounds for the probability of error for a linear classifier, where the random vectors representing the underlying classes obey the multivariate normal distribution. The expression of the error is derived in the one-dimensional space, independently of the dimensionality of the original problem. Based on the two bounds, we propose an approximating expression for the error of a generic linear classifier. In particular, we derive the corresponding bounds and the expression for approximating the error of Fisher's classifier. Our empirical results on synthetic data, including up to two-hundred-dimensional featured samples, show that the computations for the error are extremely fast and quite accurate; it differs from the actual error in at most ε=0.0184340683. The scheme has also been successfully tested on real-life data sets drawn from the UCI machine learning repository.

论文关键词:Linear discriminant analysis,Fisher's classifier,Error rate evaluation,Gaussian distributions,Curse of dimensionality

论文评审过程:Received 7 October 2004, Accepted 7 December 2004, Available online 31 March 2005.

论文官网地址:https://doi.org/10.1016/j.patcog.2004.12.002